var/home/core/zuul-output/0000755000175000017500000000000015067525005014532 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015067530771015504 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004524356515067530762017726 0ustar rootrootOct 02 16:40:31 crc systemd[1]: Starting Kubernetes Kubelet... Oct 02 16:40:31 crc restorecon[4742]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:31 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 16:40:32 crc restorecon[4742]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 16:40:32 crc restorecon[4742]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 02 16:40:33 crc kubenswrapper[4808]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 16:40:33 crc kubenswrapper[4808]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 02 16:40:33 crc kubenswrapper[4808]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 16:40:33 crc kubenswrapper[4808]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 16:40:33 crc kubenswrapper[4808]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 02 16:40:33 crc kubenswrapper[4808]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.095991 4808 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105321 4808 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105365 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105375 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105385 4808 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105400 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105409 4808 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105418 4808 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105426 4808 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105435 4808 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105443 4808 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105451 4808 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105458 4808 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105466 4808 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105474 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105482 4808 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105490 4808 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105497 4808 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105505 4808 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105512 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105521 4808 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105529 4808 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105536 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105545 4808 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105552 4808 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105560 4808 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105568 4808 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105576 4808 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105583 4808 feature_gate.go:330] unrecognized feature gate: Example Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105590 4808 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105601 4808 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105612 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105620 4808 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105628 4808 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105639 4808 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105650 4808 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105659 4808 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105671 4808 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105681 4808 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105689 4808 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105698 4808 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105707 4808 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105715 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105725 4808 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105734 4808 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105743 4808 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105751 4808 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105759 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105767 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105776 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105784 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105792 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105800 4808 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105808 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105815 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105823 4808 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105830 4808 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105838 4808 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105846 4808 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105854 4808 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105862 4808 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105870 4808 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105878 4808 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105885 4808 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105893 4808 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105901 4808 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105908 4808 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105916 4808 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105934 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105947 4808 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105956 4808 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.105966 4808 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.106958 4808 flags.go:64] FLAG: --address="0.0.0.0" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.106991 4808 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107006 4808 flags.go:64] FLAG: --anonymous-auth="true" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107017 4808 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107029 4808 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107039 4808 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107051 4808 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107062 4808 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107074 4808 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107083 4808 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107093 4808 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107103 4808 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107113 4808 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107123 4808 flags.go:64] FLAG: --cgroup-root="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107131 4808 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107140 4808 flags.go:64] FLAG: --client-ca-file="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107149 4808 flags.go:64] FLAG: --cloud-config="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107158 4808 flags.go:64] FLAG: --cloud-provider="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107167 4808 flags.go:64] FLAG: --cluster-dns="[]" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107179 4808 flags.go:64] FLAG: --cluster-domain="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107188 4808 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107198 4808 flags.go:64] FLAG: --config-dir="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107207 4808 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107216 4808 flags.go:64] FLAG: --container-log-max-files="5" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107228 4808 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107275 4808 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107284 4808 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107294 4808 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107303 4808 flags.go:64] FLAG: --contention-profiling="false" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107487 4808 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107496 4808 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107506 4808 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107515 4808 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107526 4808 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107535 4808 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107543 4808 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107552 4808 flags.go:64] FLAG: --enable-load-reader="false" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107563 4808 flags.go:64] FLAG: --enable-server="true" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107572 4808 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107583 4808 flags.go:64] FLAG: --event-burst="100" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107593 4808 flags.go:64] FLAG: --event-qps="50" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107602 4808 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107611 4808 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107629 4808 flags.go:64] FLAG: --eviction-hard="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107640 4808 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107649 4808 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107658 4808 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107668 4808 flags.go:64] FLAG: --eviction-soft="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107677 4808 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107686 4808 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107695 4808 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107706 4808 flags.go:64] FLAG: --experimental-mounter-path="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107715 4808 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107724 4808 flags.go:64] FLAG: --fail-swap-on="true" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107733 4808 flags.go:64] FLAG: --feature-gates="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107743 4808 flags.go:64] FLAG: --file-check-frequency="20s" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107752 4808 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107761 4808 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107771 4808 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107780 4808 flags.go:64] FLAG: --healthz-port="10248" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107788 4808 flags.go:64] FLAG: --help="false" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107798 4808 flags.go:64] FLAG: --hostname-override="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107806 4808 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107816 4808 flags.go:64] FLAG: --http-check-frequency="20s" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107825 4808 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107834 4808 flags.go:64] FLAG: --image-credential-provider-config="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107843 4808 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107852 4808 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107861 4808 flags.go:64] FLAG: --image-service-endpoint="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107870 4808 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107878 4808 flags.go:64] FLAG: --kube-api-burst="100" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107887 4808 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107897 4808 flags.go:64] FLAG: --kube-api-qps="50" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107907 4808 flags.go:64] FLAG: --kube-reserved="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107917 4808 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107926 4808 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107935 4808 flags.go:64] FLAG: --kubelet-cgroups="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107944 4808 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107953 4808 flags.go:64] FLAG: --lock-file="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107961 4808 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107970 4808 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107979 4808 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.107992 4808 flags.go:64] FLAG: --log-json-split-stream="false" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108001 4808 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108010 4808 flags.go:64] FLAG: --log-text-split-stream="false" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108019 4808 flags.go:64] FLAG: --logging-format="text" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108028 4808 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108037 4808 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108046 4808 flags.go:64] FLAG: --manifest-url="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108055 4808 flags.go:64] FLAG: --manifest-url-header="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108095 4808 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108106 4808 flags.go:64] FLAG: --max-open-files="1000000" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108117 4808 flags.go:64] FLAG: --max-pods="110" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108127 4808 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108136 4808 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108146 4808 flags.go:64] FLAG: --memory-manager-policy="None" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108155 4808 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108164 4808 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108174 4808 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108182 4808 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108206 4808 flags.go:64] FLAG: --node-status-max-images="50" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108214 4808 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108224 4808 flags.go:64] FLAG: --oom-score-adj="-999" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108259 4808 flags.go:64] FLAG: --pod-cidr="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108268 4808 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108285 4808 flags.go:64] FLAG: --pod-manifest-path="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108293 4808 flags.go:64] FLAG: --pod-max-pids="-1" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108303 4808 flags.go:64] FLAG: --pods-per-core="0" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108312 4808 flags.go:64] FLAG: --port="10250" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108322 4808 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108331 4808 flags.go:64] FLAG: --provider-id="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108340 4808 flags.go:64] FLAG: --qos-reserved="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108386 4808 flags.go:64] FLAG: --read-only-port="10255" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108396 4808 flags.go:64] FLAG: --register-node="true" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108406 4808 flags.go:64] FLAG: --register-schedulable="true" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108415 4808 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108439 4808 flags.go:64] FLAG: --registry-burst="10" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108448 4808 flags.go:64] FLAG: --registry-qps="5" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108457 4808 flags.go:64] FLAG: --reserved-cpus="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108466 4808 flags.go:64] FLAG: --reserved-memory="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108477 4808 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108486 4808 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108495 4808 flags.go:64] FLAG: --rotate-certificates="false" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108504 4808 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108513 4808 flags.go:64] FLAG: --runonce="false" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108522 4808 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108533 4808 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108542 4808 flags.go:64] FLAG: --seccomp-default="false" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108551 4808 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108560 4808 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108570 4808 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108579 4808 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108588 4808 flags.go:64] FLAG: --storage-driver-password="root" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108596 4808 flags.go:64] FLAG: --storage-driver-secure="false" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108605 4808 flags.go:64] FLAG: --storage-driver-table="stats" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108614 4808 flags.go:64] FLAG: --storage-driver-user="root" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108624 4808 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108634 4808 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108643 4808 flags.go:64] FLAG: --system-cgroups="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108652 4808 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108666 4808 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108675 4808 flags.go:64] FLAG: --tls-cert-file="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108683 4808 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108695 4808 flags.go:64] FLAG: --tls-min-version="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108704 4808 flags.go:64] FLAG: --tls-private-key-file="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108714 4808 flags.go:64] FLAG: --topology-manager-policy="none" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108723 4808 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108732 4808 flags.go:64] FLAG: --topology-manager-scope="container" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108741 4808 flags.go:64] FLAG: --v="2" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108752 4808 flags.go:64] FLAG: --version="false" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108763 4808 flags.go:64] FLAG: --vmodule="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108774 4808 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.108783 4808 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.108983 4808 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.108994 4808 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109002 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109010 4808 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109019 4808 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109028 4808 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109036 4808 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109044 4808 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109052 4808 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109060 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109068 4808 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109076 4808 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109083 4808 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109091 4808 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109100 4808 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109296 4808 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109304 4808 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109313 4808 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109320 4808 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109328 4808 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109336 4808 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109344 4808 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109352 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109360 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109368 4808 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109376 4808 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109384 4808 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109391 4808 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109400 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109409 4808 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109417 4808 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109425 4808 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109433 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109441 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109450 4808 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109458 4808 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109465 4808 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109474 4808 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109481 4808 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109489 4808 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109499 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109507 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109514 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109522 4808 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109529 4808 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109537 4808 feature_gate.go:330] unrecognized feature gate: Example Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109548 4808 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109559 4808 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109569 4808 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109578 4808 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109588 4808 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109599 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109609 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109617 4808 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109625 4808 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109633 4808 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109641 4808 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109648 4808 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109658 4808 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109669 4808 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109679 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109688 4808 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109696 4808 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109705 4808 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109720 4808 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109728 4808 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109737 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109745 4808 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109753 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109762 4808 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.109772 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.109797 4808 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.124824 4808 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.124881 4808 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125055 4808 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125087 4808 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125100 4808 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125111 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125127 4808 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125144 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125156 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125168 4808 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125179 4808 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125189 4808 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125200 4808 feature_gate.go:330] unrecognized feature gate: Example Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125210 4808 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125220 4808 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125293 4808 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125334 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125347 4808 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125359 4808 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125369 4808 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125379 4808 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125389 4808 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125400 4808 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125410 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125419 4808 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125430 4808 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125439 4808 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125449 4808 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125460 4808 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125470 4808 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125480 4808 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125490 4808 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125500 4808 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125511 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125521 4808 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125531 4808 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125565 4808 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125576 4808 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125586 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125597 4808 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125608 4808 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125620 4808 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125630 4808 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125674 4808 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125688 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125700 4808 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125714 4808 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125727 4808 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125738 4808 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125750 4808 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125760 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125771 4808 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125782 4808 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125792 4808 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125803 4808 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125814 4808 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125825 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125835 4808 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125849 4808 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125861 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125871 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125882 4808 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125893 4808 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125902 4808 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125913 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125923 4808 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125933 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125943 4808 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125953 4808 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125963 4808 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125973 4808 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.125982 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126013 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.126029 4808 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126459 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126481 4808 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126492 4808 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126502 4808 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126511 4808 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126522 4808 feature_gate.go:330] unrecognized feature gate: Example Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126531 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126542 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126556 4808 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126571 4808 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126583 4808 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126593 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126603 4808 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126614 4808 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126624 4808 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126634 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126644 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126654 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126664 4808 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126675 4808 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126684 4808 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126695 4808 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126705 4808 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126714 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126724 4808 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126734 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126744 4808 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126753 4808 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126764 4808 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126774 4808 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126788 4808 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126801 4808 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126812 4808 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126824 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126854 4808 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126866 4808 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126876 4808 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126888 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126898 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126909 4808 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126919 4808 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126929 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126939 4808 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126948 4808 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126958 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126971 4808 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126985 4808 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.126998 4808 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127009 4808 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127019 4808 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127029 4808 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127039 4808 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127049 4808 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127059 4808 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127075 4808 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127088 4808 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127098 4808 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127108 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127118 4808 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127128 4808 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127138 4808 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127147 4808 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127157 4808 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127167 4808 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127176 4808 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127186 4808 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127196 4808 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127206 4808 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127217 4808 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127226 4808 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.127293 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.127314 4808 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.127659 4808 server.go:940] "Client rotation is on, will bootstrap in background" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.137994 4808 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.138183 4808 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.140033 4808 server.go:997] "Starting client certificate rotation" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.140083 4808 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.141413 4808 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-29 11:50:27.924382632 +0000 UTC Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.141571 4808 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 2107h9m54.78281735s for next certificate rotation Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.179059 4808 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.183738 4808 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.202806 4808 log.go:25] "Validated CRI v1 runtime API" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.244187 4808 log.go:25] "Validated CRI v1 image API" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.246613 4808 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.252521 4808 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-02-16-35-23-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.252576 4808 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.286060 4808 manager.go:217] Machine: {Timestamp:2025-10-02 16:40:33.282818128 +0000 UTC m=+0.608347208 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654132736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:36c106c2-22f6-4c16-89bf-0f517cf57f3b BootID:ae2a1887-4119-40ce-b362-d8762c1a02f6 Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730829824 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827068416 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:3d:f3:03 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:3d:f3:03 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:8c:64:fc Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:1c:ec:bf Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:f8:a9:1f Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:d0:2d:1e Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:45:9c:24 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:2a:c5:39:3b:22:ed Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:2e:f3:65:a0:60:e4 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654132736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.286521 4808 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.286791 4808 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.289214 4808 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.289677 4808 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.289744 4808 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.290143 4808 topology_manager.go:138] "Creating topology manager with none policy" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.290171 4808 container_manager_linux.go:303] "Creating device plugin manager" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.290876 4808 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.290937 4808 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.291330 4808 state_mem.go:36] "Initialized new in-memory state store" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.291497 4808 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.296137 4808 kubelet.go:418] "Attempting to sync node with API server" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.296184 4808 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.296276 4808 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.296308 4808 kubelet.go:324] "Adding apiserver pod source" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.296331 4808 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.302868 4808 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.304155 4808 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.308989 4808 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.310798 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.310847 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.310862 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.310876 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.310899 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.310915 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.310928 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.310951 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.310967 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.310981 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.311034 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.311049 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.311124 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.311128 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 02 16:40:33 crc kubenswrapper[4808]: E1002 16:40:33.311297 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Oct 02 16:40:33 crc kubenswrapper[4808]: E1002 16:40:33.311375 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.312097 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.312884 4808 server.go:1280] "Started kubelet" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.313261 4808 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.313197 4808 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.314514 4808 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.314958 4808 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 02 16:40:33 crc systemd[1]: Started Kubernetes Kubelet. Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.316983 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.317028 4808 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.317313 4808 server.go:460] "Adding debug handlers to kubelet server" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.317423 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 13:52:39.754837221 +0000 UTC Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.317486 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 2373h12m6.437353346s for next certificate rotation Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.317752 4808 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.317782 4808 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.317934 4808 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 02 16:40:33 crc kubenswrapper[4808]: E1002 16:40:33.318041 4808 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 16:40:33 crc kubenswrapper[4808]: E1002 16:40:33.318845 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" interval="200ms" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.318972 4808 factory.go:55] Registering systemd factory Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.319657 4808 factory.go:221] Registration of the systemd container factory successfully Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.320203 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 02 16:40:33 crc kubenswrapper[4808]: E1002 16:40:33.320347 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.320916 4808 factory.go:153] Registering CRI-O factory Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.320962 4808 factory.go:221] Registration of the crio container factory successfully Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.321443 4808 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.321689 4808 factory.go:103] Registering Raw factory Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.321916 4808 manager.go:1196] Started watching for new ooms in manager Oct 02 16:40:33 crc kubenswrapper[4808]: E1002 16:40:33.320113 4808 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.13:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186aba1bdf2e5494 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-02 16:40:33.312838804 +0000 UTC m=+0.638367834,LastTimestamp:2025-10-02 16:40:33.312838804 +0000 UTC m=+0.638367834,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.324854 4808 manager.go:319] Starting recovery of all containers Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341007 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341085 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341102 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341122 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341137 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341151 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341165 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341184 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341200 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341215 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341230 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341270 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341285 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341307 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341326 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341344 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341362 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341381 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341401 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341418 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341435 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341450 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341465 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341490 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341511 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341531 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341557 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341578 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341597 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341701 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341721 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341768 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341789 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341807 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341828 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341844 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341895 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341921 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341938 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341954 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341973 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.341991 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342010 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342028 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342047 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342103 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342123 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342140 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342161 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342180 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342230 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342276 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342310 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342339 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342366 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342391 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342416 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342439 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342458 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342474 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342488 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342503 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342520 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342533 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342554 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342580 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342598 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342616 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342632 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342657 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342673 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342688 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342705 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342722 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342824 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342843 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342863 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342881 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342902 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342921 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342938 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.342954 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344706 4808 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344743 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344762 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344777 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344791 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344807 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344818 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344830 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344841 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344852 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344864 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344876 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344888 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344900 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344912 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344926 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344937 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344948 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344961 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344972 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344984 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.344996 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345010 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345029 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345043 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345056 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345071 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345085 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345101 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345114 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345127 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345142 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345156 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345169 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345183 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345197 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345220 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345356 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345374 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345387 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345438 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345450 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345469 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345482 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345502 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345520 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345539 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345555 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345582 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345599 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345616 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345635 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345652 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345669 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345686 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345705 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345719 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345737 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345757 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345776 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345804 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345820 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345835 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345849 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345863 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345878 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345902 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345916 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345930 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345943 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.345993 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346006 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346019 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346034 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346055 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346069 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346085 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346100 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346115 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346128 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346146 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346169 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346182 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346195 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346208 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346225 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346258 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346272 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346286 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346306 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346320 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346335 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346349 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346363 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346375 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346392 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346406 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346420 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346436 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346448 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346493 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346505 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346520 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346533 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346544 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346556 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346569 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346582 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346596 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346608 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346620 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346632 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346646 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346659 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346671 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346684 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346699 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346714 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346764 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346782 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346799 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346812 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346824 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346835 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346847 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346858 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346870 4808 reconstruct.go:97] "Volume reconstruction finished" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.346881 4808 reconciler.go:26] "Reconciler: start to sync state" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.364480 4808 manager.go:324] Recovery completed Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.384347 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.387548 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.387604 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.387618 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.390790 4808 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.390822 4808 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.390850 4808 state_mem.go:36] "Initialized new in-memory state store" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.392692 4808 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.394815 4808 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.394856 4808 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.394880 4808 kubelet.go:2335] "Starting kubelet main sync loop" Oct 02 16:40:33 crc kubenswrapper[4808]: E1002 16:40:33.394974 4808 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.395557 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 02 16:40:33 crc kubenswrapper[4808]: E1002 16:40:33.395614 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.411612 4808 policy_none.go:49] "None policy: Start" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.412431 4808 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.412455 4808 state_mem.go:35] "Initializing new in-memory state store" Oct 02 16:40:33 crc kubenswrapper[4808]: E1002 16:40:33.418780 4808 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.478582 4808 manager.go:334] "Starting Device Plugin manager" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.479009 4808 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.479138 4808 server.go:79] "Starting device plugin registration server" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.479945 4808 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.479978 4808 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.480194 4808 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.480358 4808 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.480373 4808 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 02 16:40:33 crc kubenswrapper[4808]: E1002 16:40:33.490437 4808 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.495756 4808 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.495871 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.497643 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.497678 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.497689 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.497817 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.497964 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.498001 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.498827 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.498883 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.499197 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.499581 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.499924 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.499970 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.499990 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.500298 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.500504 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.501211 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.501282 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.501307 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.501562 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.502269 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.502610 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.503955 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.503982 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.504019 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.504038 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.503995 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.504125 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.504298 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.504426 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.504458 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.505069 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.505094 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.505107 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.505305 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.505326 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.505337 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.505490 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.505515 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.505905 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.505934 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.505950 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.507435 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.507463 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.507480 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:33 crc kubenswrapper[4808]: E1002 16:40:33.519495 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" interval="400ms" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.548900 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.548976 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.549041 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.549114 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.549158 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.549186 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.549216 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.549284 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.549479 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.549505 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.549553 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.549580 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.549646 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.549667 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.549707 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.580390 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.581934 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.582129 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.582174 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.582199 4808 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 16:40:33 crc kubenswrapper[4808]: E1002 16:40:33.582969 4808 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.13:6443: connect: connection refused" node="crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651265 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651346 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651377 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651398 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651421 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651446 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651467 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651525 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651552 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651573 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651583 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651613 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651664 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651679 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651729 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651719 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651594 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651756 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651816 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651819 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651841 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651871 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651895 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651919 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651951 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651954 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651965 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.652018 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.651984 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.652109 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.783141 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.784387 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.784446 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.784459 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.784483 4808 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 16:40:33 crc kubenswrapper[4808]: E1002 16:40:33.785085 4808 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.13:6443: connect: connection refused" node="crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.850407 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.859373 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.881733 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.901269 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-a2d205c7119b34918e59d82287d18403ba797409332bee22351d0af602938e15 WatchSource:0}: Error finding container a2d205c7119b34918e59d82287d18403ba797409332bee22351d0af602938e15: Status 404 returned error can't find the container with id a2d205c7119b34918e59d82287d18403ba797409332bee22351d0af602938e15 Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.901766 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.903122 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-51dabfc28b34e3a4cf9a5769bed3bb05db04b232ea7ba46df1189afc6c523ade WatchSource:0}: Error finding container 51dabfc28b34e3a4cf9a5769bed3bb05db04b232ea7ba46df1189afc6c523ade: Status 404 returned error can't find the container with id 51dabfc28b34e3a4cf9a5769bed3bb05db04b232ea7ba46df1189afc6c523ade Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.906093 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-74d0327fa3a258cb431579cfbcfdc211ce4a1015913167a50e1970254ff9638c WatchSource:0}: Error finding container 74d0327fa3a258cb431579cfbcfdc211ce4a1015913167a50e1970254ff9638c: Status 404 returned error can't find the container with id 74d0327fa3a258cb431579cfbcfdc211ce4a1015913167a50e1970254ff9638c Oct 02 16:40:33 crc kubenswrapper[4808]: I1002 16:40:33.908052 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.916457 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-851b0769591e219abb63f80b8901c5b03d0a1ba6e98e8131bf04fe129b31e5d0 WatchSource:0}: Error finding container 851b0769591e219abb63f80b8901c5b03d0a1ba6e98e8131bf04fe129b31e5d0: Status 404 returned error can't find the container with id 851b0769591e219abb63f80b8901c5b03d0a1ba6e98e8131bf04fe129b31e5d0 Oct 02 16:40:33 crc kubenswrapper[4808]: E1002 16:40:33.920202 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" interval="800ms" Oct 02 16:40:33 crc kubenswrapper[4808]: W1002 16:40:33.929386 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-c5fe587a665830bb444ad47b6d89c8157e63528f8a49805f61509d214bab6deb WatchSource:0}: Error finding container c5fe587a665830bb444ad47b6d89c8157e63528f8a49805f61509d214bab6deb: Status 404 returned error can't find the container with id c5fe587a665830bb444ad47b6d89c8157e63528f8a49805f61509d214bab6deb Oct 02 16:40:34 crc kubenswrapper[4808]: I1002 16:40:34.186173 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:34 crc kubenswrapper[4808]: I1002 16:40:34.187884 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:34 crc kubenswrapper[4808]: I1002 16:40:34.188288 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:34 crc kubenswrapper[4808]: I1002 16:40:34.188308 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:34 crc kubenswrapper[4808]: I1002 16:40:34.188342 4808 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 16:40:34 crc kubenswrapper[4808]: E1002 16:40:34.188913 4808 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.13:6443: connect: connection refused" node="crc" Oct 02 16:40:34 crc kubenswrapper[4808]: W1002 16:40:34.222043 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 02 16:40:34 crc kubenswrapper[4808]: E1002 16:40:34.222184 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Oct 02 16:40:34 crc kubenswrapper[4808]: W1002 16:40:34.293338 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 02 16:40:34 crc kubenswrapper[4808]: E1002 16:40:34.293427 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Oct 02 16:40:34 crc kubenswrapper[4808]: I1002 16:40:34.316080 4808 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 02 16:40:34 crc kubenswrapper[4808]: I1002 16:40:34.399941 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"51dabfc28b34e3a4cf9a5769bed3bb05db04b232ea7ba46df1189afc6c523ade"} Oct 02 16:40:34 crc kubenswrapper[4808]: I1002 16:40:34.401667 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a2d205c7119b34918e59d82287d18403ba797409332bee22351d0af602938e15"} Oct 02 16:40:34 crc kubenswrapper[4808]: I1002 16:40:34.403075 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c5fe587a665830bb444ad47b6d89c8157e63528f8a49805f61509d214bab6deb"} Oct 02 16:40:34 crc kubenswrapper[4808]: I1002 16:40:34.404536 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"851b0769591e219abb63f80b8901c5b03d0a1ba6e98e8131bf04fe129b31e5d0"} Oct 02 16:40:34 crc kubenswrapper[4808]: I1002 16:40:34.405831 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"74d0327fa3a258cb431579cfbcfdc211ce4a1015913167a50e1970254ff9638c"} Oct 02 16:40:34 crc kubenswrapper[4808]: W1002 16:40:34.454095 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 02 16:40:34 crc kubenswrapper[4808]: E1002 16:40:34.454207 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Oct 02 16:40:34 crc kubenswrapper[4808]: W1002 16:40:34.531213 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 02 16:40:34 crc kubenswrapper[4808]: E1002 16:40:34.531302 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Oct 02 16:40:34 crc kubenswrapper[4808]: E1002 16:40:34.721841 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" interval="1.6s" Oct 02 16:40:34 crc kubenswrapper[4808]: I1002 16:40:34.989026 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:34 crc kubenswrapper[4808]: I1002 16:40:34.990790 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:34 crc kubenswrapper[4808]: I1002 16:40:34.990836 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:34 crc kubenswrapper[4808]: I1002 16:40:34.990851 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:34 crc kubenswrapper[4808]: I1002 16:40:34.990882 4808 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 16:40:34 crc kubenswrapper[4808]: E1002 16:40:34.991612 4808 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.13:6443: connect: connection refused" node="crc" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.316814 4808 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.408528 4808 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615" exitCode=0 Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.408606 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615"} Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.408688 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.409521 4808 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="0a254a3cf9d11b54b681fc4cda92b06f88ff70b87906c439cdaecc58c7b0738f" exitCode=0 Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.409552 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"0a254a3cf9d11b54b681fc4cda92b06f88ff70b87906c439cdaecc58c7b0738f"} Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.409575 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.410197 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.410228 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.410264 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.410766 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.410816 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.410834 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.411402 4808 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5" exitCode=0 Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.411466 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5"} Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.411491 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.415600 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.415638 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.415647 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.417281 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.417274 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00"} Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.417388 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c"} Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.417412 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493"} Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.417429 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9"} Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.418172 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.418200 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.418214 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.418866 4808 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2" exitCode=0 Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.418906 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2"} Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.418945 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.419985 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.420043 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.420062 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.422035 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.423003 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.423417 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:35 crc kubenswrapper[4808]: I1002 16:40:35.423485 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:35 crc kubenswrapper[4808]: E1002 16:40:35.682562 4808 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.13:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186aba1bdf2e5494 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-02 16:40:33.312838804 +0000 UTC m=+0.638367834,LastTimestamp:2025-10-02 16:40:33.312838804 +0000 UTC m=+0.638367834,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.135326 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.315884 4808 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 02 16:40:36 crc kubenswrapper[4808]: E1002 16:40:36.323299 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" interval="3.2s" Oct 02 16:40:36 crc kubenswrapper[4808]: W1002 16:40:36.342746 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 02 16:40:36 crc kubenswrapper[4808]: E1002 16:40:36.342839 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.429089 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a"} Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.429141 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459"} Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.429157 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2"} Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.429169 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0"} Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.431152 4808 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50" exitCode=0 Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.431181 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50"} Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.431336 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.432365 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.432407 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.432423 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.432610 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"5fe995b6f7e46579fae83eb2918d20e9801ed0f52df4753ad1871c41b9e11abc"} Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.432665 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.434024 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.434041 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.434050 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.438411 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.439137 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.439619 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5dc88fc50e88cd95a703f74122351a86eee2dc44b402e48893690ecd79c9e5b3"} Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.439666 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e28b9ee75280e7e9c44e99c1b82c9760990628488cb359247df6483530fa09ef"} Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.439684 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"07b9557dcbd29f15e9ae34cc712160a7f5f34330deaa1c08774efd54e68c4ac4"} Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.440171 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.440326 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.440566 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.440628 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.440972 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.441000 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.591712 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.592926 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.592973 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.592989 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:36 crc kubenswrapper[4808]: I1002 16:40:36.593023 4808 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 16:40:36 crc kubenswrapper[4808]: E1002 16:40:36.593466 4808 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.13:6443: connect: connection refused" node="crc" Oct 02 16:40:36 crc kubenswrapper[4808]: W1002 16:40:36.799165 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 02 16:40:36 crc kubenswrapper[4808]: E1002 16:40:36.799305 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.443773 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2"} Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.443940 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.445443 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.445503 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.445538 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.446378 4808 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30" exitCode=0 Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.446594 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.446655 4808 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.446679 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.446704 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.446793 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30"} Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.446596 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.448964 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.448986 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.449015 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.448964 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.449033 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.449019 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.449058 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.449039 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.449191 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.449218 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.449294 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:37 crc kubenswrapper[4808]: I1002 16:40:37.449319 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:38 crc kubenswrapper[4808]: I1002 16:40:38.453595 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:38 crc kubenswrapper[4808]: I1002 16:40:38.453595 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0"} Oct 02 16:40:38 crc kubenswrapper[4808]: I1002 16:40:38.453674 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde"} Oct 02 16:40:38 crc kubenswrapper[4808]: I1002 16:40:38.453692 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07"} Oct 02 16:40:38 crc kubenswrapper[4808]: I1002 16:40:38.453705 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612"} Oct 02 16:40:38 crc kubenswrapper[4808]: I1002 16:40:38.453722 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 16:40:38 crc kubenswrapper[4808]: I1002 16:40:38.454483 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:38 crc kubenswrapper[4808]: I1002 16:40:38.454529 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:38 crc kubenswrapper[4808]: I1002 16:40:38.454544 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.135960 4808 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.136094 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.463257 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17"} Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.463461 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.463476 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.465198 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.465284 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.465298 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.465892 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.465963 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.465989 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.793698 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.795557 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.795631 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.795649 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.795749 4808 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.959635 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.959869 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.961684 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.961777 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:39 crc kubenswrapper[4808]: I1002 16:40:39.961815 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:40 crc kubenswrapper[4808]: I1002 16:40:40.013482 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 02 16:40:40 crc kubenswrapper[4808]: I1002 16:40:40.465576 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:40 crc kubenswrapper[4808]: I1002 16:40:40.466873 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:40 crc kubenswrapper[4808]: I1002 16:40:40.466943 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:40 crc kubenswrapper[4808]: I1002 16:40:40.466965 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:41 crc kubenswrapper[4808]: I1002 16:40:41.033194 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 16:40:41 crc kubenswrapper[4808]: I1002 16:40:41.033504 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:41 crc kubenswrapper[4808]: I1002 16:40:41.035160 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:41 crc kubenswrapper[4808]: I1002 16:40:41.035271 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:41 crc kubenswrapper[4808]: I1002 16:40:41.035291 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:41 crc kubenswrapper[4808]: I1002 16:40:41.469107 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:41 crc kubenswrapper[4808]: I1002 16:40:41.470565 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:41 crc kubenswrapper[4808]: I1002 16:40:41.470619 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:41 crc kubenswrapper[4808]: I1002 16:40:41.470640 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:42 crc kubenswrapper[4808]: I1002 16:40:42.018690 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 16:40:42 crc kubenswrapper[4808]: I1002 16:40:42.018992 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 16:40:42 crc kubenswrapper[4808]: I1002 16:40:42.019092 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:42 crc kubenswrapper[4808]: I1002 16:40:42.020586 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:42 crc kubenswrapper[4808]: I1002 16:40:42.020630 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:42 crc kubenswrapper[4808]: I1002 16:40:42.020640 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:42 crc kubenswrapper[4808]: I1002 16:40:42.472060 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:42 crc kubenswrapper[4808]: I1002 16:40:42.473525 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:42 crc kubenswrapper[4808]: I1002 16:40:42.473570 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:42 crc kubenswrapper[4808]: I1002 16:40:42.473581 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:42 crc kubenswrapper[4808]: I1002 16:40:42.728105 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 02 16:40:42 crc kubenswrapper[4808]: I1002 16:40:42.728402 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:42 crc kubenswrapper[4808]: I1002 16:40:42.730012 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:42 crc kubenswrapper[4808]: I1002 16:40:42.730085 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:42 crc kubenswrapper[4808]: I1002 16:40:42.730104 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:43 crc kubenswrapper[4808]: E1002 16:40:43.490692 4808 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 16:40:44 crc kubenswrapper[4808]: I1002 16:40:44.161499 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 16:40:44 crc kubenswrapper[4808]: I1002 16:40:44.161712 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:44 crc kubenswrapper[4808]: I1002 16:40:44.163170 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:44 crc kubenswrapper[4808]: I1002 16:40:44.163216 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:44 crc kubenswrapper[4808]: I1002 16:40:44.163270 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:44 crc kubenswrapper[4808]: I1002 16:40:44.168875 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 16:40:44 crc kubenswrapper[4808]: I1002 16:40:44.478177 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:44 crc kubenswrapper[4808]: I1002 16:40:44.479924 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:44 crc kubenswrapper[4808]: I1002 16:40:44.480178 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:44 crc kubenswrapper[4808]: I1002 16:40:44.480378 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:44 crc kubenswrapper[4808]: I1002 16:40:44.485584 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 16:40:44 crc kubenswrapper[4808]: I1002 16:40:44.605278 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 16:40:45 crc kubenswrapper[4808]: I1002 16:40:45.480900 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:45 crc kubenswrapper[4808]: I1002 16:40:45.482948 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:45 crc kubenswrapper[4808]: I1002 16:40:45.483020 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:45 crc kubenswrapper[4808]: I1002 16:40:45.483040 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:46 crc kubenswrapper[4808]: I1002 16:40:46.484547 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:46 crc kubenswrapper[4808]: I1002 16:40:46.489084 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:46 crc kubenswrapper[4808]: I1002 16:40:46.489150 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:46 crc kubenswrapper[4808]: I1002 16:40:46.489171 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:47 crc kubenswrapper[4808]: W1002 16:40:47.070878 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 02 16:40:47 crc kubenswrapper[4808]: I1002 16:40:47.071024 4808 trace.go:236] Trace[2016379963]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 16:40:37.069) (total time: 10001ms): Oct 02 16:40:47 crc kubenswrapper[4808]: Trace[2016379963]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (16:40:47.070) Oct 02 16:40:47 crc kubenswrapper[4808]: Trace[2016379963]: [10.001663696s] [10.001663696s] END Oct 02 16:40:47 crc kubenswrapper[4808]: E1002 16:40:47.071054 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 02 16:40:47 crc kubenswrapper[4808]: I1002 16:40:47.214669 4808 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 02 16:40:47 crc kubenswrapper[4808]: I1002 16:40:47.214755 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 02 16:40:47 crc kubenswrapper[4808]: I1002 16:40:47.219256 4808 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 02 16:40:47 crc kubenswrapper[4808]: I1002 16:40:47.219328 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 02 16:40:49 crc kubenswrapper[4808]: I1002 16:40:49.136578 4808 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 16:40:49 crc kubenswrapper[4808]: I1002 16:40:49.137041 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 02 16:40:51 crc kubenswrapper[4808]: I1002 16:40:51.180718 4808 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.023682 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.023887 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.025651 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.025698 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.025713 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.027763 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.168550 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.170405 4808 trace.go:236] Trace[1370702544]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 16:40:37.392) (total time: 14777ms): Oct 02 16:40:52 crc kubenswrapper[4808]: Trace[1370702544]: ---"Objects listed" error: 14777ms (16:40:52.170) Oct 02 16:40:52 crc kubenswrapper[4808]: Trace[1370702544]: [14.777361901s] [14.777361901s] END Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.170436 4808 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.170773 4808 trace.go:236] Trace[1918225829]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 16:40:42.065) (total time: 10105ms): Oct 02 16:40:52 crc kubenswrapper[4808]: Trace[1918225829]: ---"Objects listed" error: 10105ms (16:40:52.170) Oct 02 16:40:52 crc kubenswrapper[4808]: Trace[1918225829]: [10.105516416s] [10.105516416s] END Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.170786 4808 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.171083 4808 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.171781 4808 trace.go:236] Trace[139786345]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 16:40:40.097) (total time: 12074ms): Oct 02 16:40:52 crc kubenswrapper[4808]: Trace[139786345]: ---"Objects listed" error: 12074ms (16:40:52.171) Oct 02 16:40:52 crc kubenswrapper[4808]: Trace[139786345]: [12.074712178s] [12.074712178s] END Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.171808 4808 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.171843 4808 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.192992 4808 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:45850->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.193085 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:45850->192.168.126.11:17697: read: connection reset by peer" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.193140 4808 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:52070->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.193266 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:52070->192.168.126.11:17697: read: connection reset by peer" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.306593 4808 apiserver.go:52] "Watching apiserver" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.310163 4808 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.310527 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.310969 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.311038 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.311134 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.311271 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.311507 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.311532 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.311556 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.311617 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.311690 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.312672 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.313430 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.313657 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.313933 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.314067 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.314431 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.314968 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.315144 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.316659 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.319255 4808 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.342126 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.354721 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.369730 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372340 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372384 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372404 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372424 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372442 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372457 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372472 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372490 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372510 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372527 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372591 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372635 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372651 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372669 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372704 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372722 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372743 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372812 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372829 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372848 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372865 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372917 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372943 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372955 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.373001 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.373019 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.373036 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.372978 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.373054 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.373073 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.373089 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.373105 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.373125 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.373141 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.373109 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.373158 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.373975 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.373994 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374121 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374205 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374263 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374313 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374301 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374360 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374359 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374415 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374467 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374499 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374542 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374581 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374590 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374613 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374703 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374746 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374764 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374847 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374860 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.374896 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:40:52.874868899 +0000 UTC m=+20.200397899 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374940 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374957 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374986 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375024 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375027 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375107 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.374990 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375182 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375226 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375272 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375290 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375317 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375353 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375464 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375496 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375580 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375622 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375650 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375673 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375705 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376001 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376050 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376073 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376099 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376127 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376150 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376181 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376217 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376259 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376288 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376316 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376378 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376406 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376550 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376583 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376606 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376631 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376657 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376676 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376708 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376740 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376778 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376822 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376853 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376884 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376910 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376941 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376973 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.377002 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.377038 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.377071 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.377099 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.377138 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.377169 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.377196 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.377215 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.377431 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.377457 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.377474 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.377497 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.377520 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.377541 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.382950 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.382999 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.383043 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.383087 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.383311 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.383357 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.383392 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.383420 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.383476 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.383512 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.383545 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.383576 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.383605 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.383661 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.383700 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.383732 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.383768 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.383924 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.383989 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384040 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384105 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384151 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384212 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384296 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384336 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384401 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384430 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384458 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384517 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384569 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384615 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384639 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384664 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384879 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384909 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384940 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384995 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.385053 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388005 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388057 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388083 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388102 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388132 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388157 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388187 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388208 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388246 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388279 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388300 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388334 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388353 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388374 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388395 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388420 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388461 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388483 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388503 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388520 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.389056 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375312 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.391982 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.392351 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.392645 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.392772 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.392851 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.392909 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.392900 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375357 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375412 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375532 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375705 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375823 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.375848 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.376774 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.377136 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.377340 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.377602 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.377721 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.378292 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.378320 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.378692 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.378992 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.379021 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.379312 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.382712 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.382938 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.383404 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.383846 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.383890 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384483 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384544 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384700 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.384877 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.385107 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.385169 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.385297 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.385534 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.385643 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.385695 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.385830 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.386109 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.386261 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.386391 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.386725 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.386772 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.386783 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.387380 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.387506 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.387569 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.387593 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.387970 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388342 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388223 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388505 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388701 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388797 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388906 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388947 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.388933 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.389436 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.389846 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.390616 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.390916 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.391189 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.391348 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.389042 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.391980 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.392103 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.392161 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.393018 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.393071 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.393289 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.393366 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.393722 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.393795 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.392806 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.394415 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.394487 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.394634 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.394726 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.395365 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.395503 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.395649 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.395857 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.396154 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.396408 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.396871 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.397305 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.398098 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.399146 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.399283 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.399672 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.400403 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.400462 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.400642 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.400964 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.400992 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.401281 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.401556 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.401569 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.401703 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.401894 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.401907 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.401942 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.402116 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.402320 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.402478 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.402539 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.402612 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.402791 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.402819 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.402831 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.402916 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.403003 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.403039 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.392983 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.403289 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.403297 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.403300 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.403328 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.403409 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.403472 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.403872 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.403907 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.403929 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.403990 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.404016 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.404035 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.406220 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.408205 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.408316 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.408515 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.408689 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.408707 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.408959 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409007 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.406419 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.407400 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.408776 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.408825 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-7v5x7"] Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409095 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409219 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409346 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409385 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409412 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409435 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409465 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409497 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409519 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409543 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409573 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409620 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409648 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409680 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409710 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409741 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409756 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-7v5x7" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409773 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409837 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409879 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409909 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409935 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409964 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409990 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410017 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410075 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410117 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410154 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410191 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410213 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410249 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410275 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410409 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410424 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410435 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410447 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410458 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410468 4808 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410480 4808 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410491 4808 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410502 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410514 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410525 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410535 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410545 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410554 4808 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410563 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410575 4808 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410585 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410596 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410606 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410616 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410704 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410727 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410745 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410758 4808 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410772 4808 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410787 4808 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410800 4808 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410813 4808 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410824 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410833 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410844 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410853 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410864 4808 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410873 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410882 4808 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410891 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410900 4808 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410909 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410921 4808 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410930 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410941 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410951 4808 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410960 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410969 4808 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410979 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410988 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.410999 4808 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411008 4808 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411017 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411026 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411035 4808 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411045 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411055 4808 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411063 4808 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411073 4808 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411082 4808 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411091 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411100 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411109 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411117 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411126 4808 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411134 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411145 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411157 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411172 4808 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411185 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411199 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411212 4808 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411224 4808 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411251 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411259 4808 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411268 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411277 4808 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411288 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411297 4808 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411310 4808 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411322 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411334 4808 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411347 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411362 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411376 4808 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411389 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411400 4808 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411414 4808 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411429 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411474 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411488 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411500 4808 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411524 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411538 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411550 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411563 4808 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411574 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411585 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411597 4808 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411608 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411620 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411632 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411644 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411656 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411668 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411679 4808 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411700 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411713 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411724 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411736 4808 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411749 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411761 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411775 4808 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411787 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411801 4808 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411811 4808 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411821 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411833 4808 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411844 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411856 4808 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411868 4808 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411883 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411894 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411905 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411917 4808 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411929 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411941 4808 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411954 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411965 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411977 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.411990 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.412003 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.412015 4808 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.412026 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.412040 4808 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.412051 4808 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.412062 4808 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.412074 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.412085 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.412097 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.412111 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.412121 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.412134 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.412914 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.414083 4808 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409339 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409551 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409609 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409894 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.409970 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.412118 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.412226 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.412380 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.415833 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.412575 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.412583 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.413043 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.413355 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.413500 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.414029 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.415942 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.414166 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.414222 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.414341 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.416096 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 16:40:52.916073805 +0000 UTC m=+20.241602805 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.416099 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.414497 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.414648 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.415135 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.416170 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.416292 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.416308 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.415158 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.415320 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.415643 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.415740 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.416486 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 16:40:52.916175557 +0000 UTC m=+20.241704547 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.416524 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.416801 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.417131 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.417164 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.417427 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.417429 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.417618 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.417912 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.418640 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.418783 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.419057 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.419318 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.419392 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.421090 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.421279 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.421388 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.421574 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.421691 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.427133 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.427510 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.428379 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.428549 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.432701 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.432985 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.434001 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.436549 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.436651 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.436677 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.436696 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.436760 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.436782 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 16:40:52.936757025 +0000 UTC m=+20.262286225 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.436901 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.436935 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.436953 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.437030 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 16:40:52.937001451 +0000 UTC m=+20.262530661 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.437794 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.437913 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.438798 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.439732 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.441065 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.441153 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.449891 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.454612 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.454864 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.457928 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.470021 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.471403 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.477362 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.480191 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.483903 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.490738 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.503106 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.506425 4808 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2" exitCode=255 Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.506478 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2"} Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.509354 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513450 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513506 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tptqw\" (UniqueName: \"kubernetes.io/projected/78096007-76e2-47b5-a1c1-549725c4c9a7-kube-api-access-tptqw\") pod \"node-resolver-7v5x7\" (UID: \"78096007-76e2-47b5-a1c1-549725c4c9a7\") " pod="openshift-dns/node-resolver-7v5x7" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513528 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513563 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/78096007-76e2-47b5-a1c1-549725c4c9a7-hosts-file\") pod \"node-resolver-7v5x7\" (UID: \"78096007-76e2-47b5-a1c1-549725c4c9a7\") " pod="openshift-dns/node-resolver-7v5x7" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513593 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513607 4808 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513618 4808 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513627 4808 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513652 4808 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513661 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513671 4808 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513681 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513690 4808 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513699 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513713 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513723 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513733 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513742 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513753 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513764 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513774 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513784 4808 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513798 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513809 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513821 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513832 4808 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513843 4808 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513853 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513863 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513873 4808 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513883 4808 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513893 4808 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513906 4808 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513916 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513929 4808 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513941 4808 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513953 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513965 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514180 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514247 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.513977 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514299 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514350 4808 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514365 4808 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514379 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514392 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514407 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514418 4808 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514433 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514446 4808 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514458 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514471 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514483 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514495 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514506 4808 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514518 4808 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514530 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514542 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514553 4808 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514578 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.514589 4808 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.524191 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.526671 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.526748 4808 scope.go:117] "RemoveContainer" containerID="78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.535633 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.551990 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.565013 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.577026 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.590253 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.615600 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tptqw\" (UniqueName: \"kubernetes.io/projected/78096007-76e2-47b5-a1c1-549725c4c9a7-kube-api-access-tptqw\") pod \"node-resolver-7v5x7\" (UID: \"78096007-76e2-47b5-a1c1-549725c4c9a7\") " pod="openshift-dns/node-resolver-7v5x7" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.615647 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/78096007-76e2-47b5-a1c1-549725c4c9a7-hosts-file\") pod \"node-resolver-7v5x7\" (UID: \"78096007-76e2-47b5-a1c1-549725c4c9a7\") " pod="openshift-dns/node-resolver-7v5x7" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.615736 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/78096007-76e2-47b5-a1c1-549725c4c9a7-hosts-file\") pod \"node-resolver-7v5x7\" (UID: \"78096007-76e2-47b5-a1c1-549725c4c9a7\") " pod="openshift-dns/node-resolver-7v5x7" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.615851 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.624451 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.632414 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.640353 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.640832 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tptqw\" (UniqueName: \"kubernetes.io/projected/78096007-76e2-47b5-a1c1-549725c4c9a7-kube-api-access-tptqw\") pod \"node-resolver-7v5x7\" (UID: \"78096007-76e2-47b5-a1c1-549725c4c9a7\") " pod="openshift-dns/node-resolver-7v5x7" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.644774 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.662356 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: W1002 16:40:52.668713 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-9d3a35055af7aad2573f8beb1a87543bb9b098b791a6199a83e164bf174af43f WatchSource:0}: Error finding container 9d3a35055af7aad2573f8beb1a87543bb9b098b791a6199a83e164bf174af43f: Status 404 returned error can't find the container with id 9d3a35055af7aad2573f8beb1a87543bb9b098b791a6199a83e164bf174af43f Oct 02 16:40:52 crc kubenswrapper[4808]: W1002 16:40:52.670657 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-e3915de4d0718b36d8287278672aa482d71b825199dbbbde5457b7882275020e WatchSource:0}: Error finding container e3915de4d0718b36d8287278672aa482d71b825199dbbbde5457b7882275020e: Status 404 returned error can't find the container with id e3915de4d0718b36d8287278672aa482d71b825199dbbbde5457b7882275020e Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.675012 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: W1002 16:40:52.678426 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-d2e282b5ec28190227fee102f69a68c35ec2383fb4f32be2a0537061b546195a WatchSource:0}: Error finding container d2e282b5ec28190227fee102f69a68c35ec2383fb4f32be2a0537061b546195a: Status 404 returned error can't find the container with id d2e282b5ec28190227fee102f69a68c35ec2383fb4f32be2a0537061b546195a Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.686122 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.702713 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.758810 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-7v5x7" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.775624 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 02 16:40:52 crc kubenswrapper[4808]: W1002 16:40:52.794093 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78096007_76e2_47b5_a1c1_549725c4c9a7.slice/crio-011ffc5aacb5fb3c22309b95f5274c8f86801cc356e6a181a6b34be471b4fe48 WatchSource:0}: Error finding container 011ffc5aacb5fb3c22309b95f5274c8f86801cc356e6a181a6b34be471b4fe48: Status 404 returned error can't find the container with id 011ffc5aacb5fb3c22309b95f5274c8f86801cc356e6a181a6b34be471b4fe48 Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.800311 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.801396 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.802778 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.817108 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.831344 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.842873 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.859635 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.873441 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.883492 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.892850 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.913181 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.918057 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.918118 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.918140 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.918183 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:40:53.918166972 +0000 UTC m=+21.243695972 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.918246 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.918255 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.918275 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 16:40:53.918266354 +0000 UTC m=+21.243795354 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 16:40:52 crc kubenswrapper[4808]: E1002 16:40:52.918288 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 16:40:53.918282675 +0000 UTC m=+21.243811675 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.926372 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.939080 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.955615 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.968436 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.980322 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:52 crc kubenswrapper[4808]: I1002 16:40:52.992549 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.004588 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.013457 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.018814 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.018852 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:40:53 crc kubenswrapper[4808]: E1002 16:40:53.018964 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 16:40:53 crc kubenswrapper[4808]: E1002 16:40:53.018979 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 16:40:53 crc kubenswrapper[4808]: E1002 16:40:53.018990 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:40:53 crc kubenswrapper[4808]: E1002 16:40:53.019038 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 16:40:54.019024566 +0000 UTC m=+21.344553566 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:40:53 crc kubenswrapper[4808]: E1002 16:40:53.019104 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 16:40:53 crc kubenswrapper[4808]: E1002 16:40:53.019141 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 16:40:53 crc kubenswrapper[4808]: E1002 16:40:53.019156 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:40:53 crc kubenswrapper[4808]: E1002 16:40:53.019252 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 16:40:54.019218041 +0000 UTC m=+21.344747081 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.410564 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.418620 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.420725 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.428478 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.429354 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.430129 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.430829 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.431077 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.431722 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.443904 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.444116 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.445197 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.445914 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.446683 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.447616 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.448347 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.449004 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.449696 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.450426 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.451260 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.451815 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.452625 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.456177 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.456263 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.456921 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.457662 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.458731 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.459688 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.460923 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.461818 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.463140 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.463789 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.465199 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.465993 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.466775 4808 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.466915 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.468596 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.469761 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.470516 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.470979 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.472143 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.472791 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.473350 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.474161 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.476399 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.476873 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.477824 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.478908 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.479540 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.480402 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.480968 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.481919 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.482715 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.483591 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.484127 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.484614 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.485570 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.486189 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.487316 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.494003 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.508653 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.513540 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-7v5x7" event={"ID":"78096007-76e2-47b5-a1c1-549725c4c9a7","Type":"ContainerStarted","Data":"9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742"} Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.513591 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-7v5x7" event={"ID":"78096007-76e2-47b5-a1c1-549725c4c9a7","Type":"ContainerStarted","Data":"011ffc5aacb5fb3c22309b95f5274c8f86801cc356e6a181a6b34be471b4fe48"} Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.515889 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7"} Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.515947 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5"} Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.515961 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d2e282b5ec28190227fee102f69a68c35ec2383fb4f32be2a0537061b546195a"} Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.516863 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e3915de4d0718b36d8287278672aa482d71b825199dbbbde5457b7882275020e"} Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.518122 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706"} Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.518174 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"9d3a35055af7aad2573f8beb1a87543bb9b098b791a6199a83e164bf174af43f"} Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.519743 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.521556 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf"} Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.524195 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.532069 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.552218 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.566813 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.594508 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.607658 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.636251 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.653753 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.679161 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.704912 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.721428 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.925981 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.926040 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:40:53 crc kubenswrapper[4808]: I1002 16:40:53.926064 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:40:53 crc kubenswrapper[4808]: E1002 16:40:53.926169 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 16:40:53 crc kubenswrapper[4808]: E1002 16:40:53.926220 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:40:55.926178362 +0000 UTC m=+23.251707352 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:40:53 crc kubenswrapper[4808]: E1002 16:40:53.926259 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 16:40:53 crc kubenswrapper[4808]: E1002 16:40:53.926287 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 16:40:55.926274954 +0000 UTC m=+23.251804204 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 16:40:53 crc kubenswrapper[4808]: E1002 16:40:53.926405 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 16:40:55.926377517 +0000 UTC m=+23.251906517 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.026863 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.026911 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:40:54 crc kubenswrapper[4808]: E1002 16:40:54.027022 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 16:40:54 crc kubenswrapper[4808]: E1002 16:40:54.027039 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 16:40:54 crc kubenswrapper[4808]: E1002 16:40:54.027052 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:40:54 crc kubenswrapper[4808]: E1002 16:40:54.027106 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 16:40:56.027092358 +0000 UTC m=+23.352621358 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:40:54 crc kubenswrapper[4808]: E1002 16:40:54.027103 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 16:40:54 crc kubenswrapper[4808]: E1002 16:40:54.027143 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 16:40:54 crc kubenswrapper[4808]: E1002 16:40:54.027154 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:40:54 crc kubenswrapper[4808]: E1002 16:40:54.027208 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 16:40:56.02719226 +0000 UTC m=+23.352721260 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.299967 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-ppf7n"] Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.300469 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-ppf7n" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.305759 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.306342 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.306366 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.306381 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.324942 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:54Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.338375 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:54Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.355673 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:54Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.375495 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:54Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.391963 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:54Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.395044 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.395044 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:40:54 crc kubenswrapper[4808]: E1002 16:40:54.395164 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:40:54 crc kubenswrapper[4808]: E1002 16:40:54.395248 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.395050 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:40:54 crc kubenswrapper[4808]: E1002 16:40:54.395326 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.406477 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:54Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.426753 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:54Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.430759 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/96534061-3b16-458d-813f-507d5d0acbcc-host\") pod \"node-ca-ppf7n\" (UID: \"96534061-3b16-458d-813f-507d5d0acbcc\") " pod="openshift-image-registry/node-ca-ppf7n" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.431215 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/96534061-3b16-458d-813f-507d5d0acbcc-serviceca\") pod \"node-ca-ppf7n\" (UID: \"96534061-3b16-458d-813f-507d5d0acbcc\") " pod="openshift-image-registry/node-ca-ppf7n" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.431311 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h9sl\" (UniqueName: \"kubernetes.io/projected/96534061-3b16-458d-813f-507d5d0acbcc-kube-api-access-8h9sl\") pod \"node-ca-ppf7n\" (UID: \"96534061-3b16-458d-813f-507d5d0acbcc\") " pod="openshift-image-registry/node-ca-ppf7n" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.447127 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:54Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.461056 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:54Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.471177 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:54Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.524276 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.532405 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/96534061-3b16-458d-813f-507d5d0acbcc-host\") pod \"node-ca-ppf7n\" (UID: \"96534061-3b16-458d-813f-507d5d0acbcc\") " pod="openshift-image-registry/node-ca-ppf7n" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.532457 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/96534061-3b16-458d-813f-507d5d0acbcc-serviceca\") pod \"node-ca-ppf7n\" (UID: \"96534061-3b16-458d-813f-507d5d0acbcc\") " pod="openshift-image-registry/node-ca-ppf7n" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.532549 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/96534061-3b16-458d-813f-507d5d0acbcc-host\") pod \"node-ca-ppf7n\" (UID: \"96534061-3b16-458d-813f-507d5d0acbcc\") " pod="openshift-image-registry/node-ca-ppf7n" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.533810 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/96534061-3b16-458d-813f-507d5d0acbcc-serviceca\") pod \"node-ca-ppf7n\" (UID: \"96534061-3b16-458d-813f-507d5d0acbcc\") " pod="openshift-image-registry/node-ca-ppf7n" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.534003 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h9sl\" (UniqueName: \"kubernetes.io/projected/96534061-3b16-458d-813f-507d5d0acbcc-kube-api-access-8h9sl\") pod \"node-ca-ppf7n\" (UID: \"96534061-3b16-458d-813f-507d5d0acbcc\") " pod="openshift-image-registry/node-ca-ppf7n" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.552264 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h9sl\" (UniqueName: \"kubernetes.io/projected/96534061-3b16-458d-813f-507d5d0acbcc-kube-api-access-8h9sl\") pod \"node-ca-ppf7n\" (UID: \"96534061-3b16-458d-813f-507d5d0acbcc\") " pod="openshift-image-registry/node-ca-ppf7n" Oct 02 16:40:54 crc kubenswrapper[4808]: I1002 16:40:54.620072 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-ppf7n" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.091529 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-7d9g4"] Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.091867 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.092606 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-75l8m"] Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.093371 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.093874 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.093910 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.093997 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-7z66r"] Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.094016 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.094072 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.094412 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.097215 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.097354 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.097541 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.098546 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.098554 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.098656 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.099817 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.100160 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.114056 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.125009 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.136920 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.145804 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.156619 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.167869 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.185148 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.196399 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.209992 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.218227 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.234366 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240052 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-multus-conf-dir\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240095 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5716183c-6301-4312-923c-e34254575a82-tuning-conf-dir\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240115 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8871f396-b980-46a8-9477-b3be4889638f-cni-binary-copy\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240142 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqqm5\" (UniqueName: \"kubernetes.io/projected/5716183c-6301-4312-923c-e34254575a82-kube-api-access-qqqm5\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240164 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pw52\" (UniqueName: \"kubernetes.io/projected/f83ce425-101d-4489-94a4-5c256eb29328-kube-api-access-8pw52\") pod \"machine-config-daemon-7z66r\" (UID: \"f83ce425-101d-4489-94a4-5c256eb29328\") " pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240181 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-multus-socket-dir-parent\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240200 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5716183c-6301-4312-923c-e34254575a82-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240242 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-host-run-multus-certs\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240261 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-os-release\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240293 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f83ce425-101d-4489-94a4-5c256eb29328-rootfs\") pod \"machine-config-daemon-7z66r\" (UID: \"f83ce425-101d-4489-94a4-5c256eb29328\") " pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240321 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5716183c-6301-4312-923c-e34254575a82-cni-binary-copy\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240336 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-cnibin\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240352 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-host-var-lib-cni-bin\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240372 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8871f396-b980-46a8-9477-b3be4889638f-multus-daemon-config\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240390 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmcjn\" (UniqueName: \"kubernetes.io/projected/8871f396-b980-46a8-9477-b3be4889638f-kube-api-access-cmcjn\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240407 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-host-run-netns\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240422 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-hostroot\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240522 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-host-var-lib-cni-multus\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240592 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5716183c-6301-4312-923c-e34254575a82-cnibin\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240618 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5716183c-6301-4312-923c-e34254575a82-os-release\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240638 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-host-var-lib-kubelet\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240662 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-etc-kubernetes\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240776 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-multus-cni-dir\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240834 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5716183c-6301-4312-923c-e34254575a82-system-cni-dir\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240867 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f83ce425-101d-4489-94a4-5c256eb29328-mcd-auth-proxy-config\") pod \"machine-config-daemon-7z66r\" (UID: \"f83ce425-101d-4489-94a4-5c256eb29328\") " pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240894 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-host-run-k8s-cni-cncf-io\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240936 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f83ce425-101d-4489-94a4-5c256eb29328-proxy-tls\") pod \"machine-config-daemon-7z66r\" (UID: \"f83ce425-101d-4489-94a4-5c256eb29328\") " pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.240960 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-system-cni-dir\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.244990 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.259148 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.278718 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.296446 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.312151 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.326780 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.338920 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.341603 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-multus-cni-dir\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.341643 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-etc-kubernetes\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.341671 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f83ce425-101d-4489-94a4-5c256eb29328-mcd-auth-proxy-config\") pod \"machine-config-daemon-7z66r\" (UID: \"f83ce425-101d-4489-94a4-5c256eb29328\") " pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.341696 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5716183c-6301-4312-923c-e34254575a82-system-cni-dir\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.341729 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f83ce425-101d-4489-94a4-5c256eb29328-proxy-tls\") pod \"machine-config-daemon-7z66r\" (UID: \"f83ce425-101d-4489-94a4-5c256eb29328\") " pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.341751 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-host-run-k8s-cni-cncf-io\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.341773 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-system-cni-dir\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.341794 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5716183c-6301-4312-923c-e34254575a82-tuning-conf-dir\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.341818 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8871f396-b980-46a8-9477-b3be4889638f-cni-binary-copy\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.341842 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-multus-conf-dir\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.341882 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pw52\" (UniqueName: \"kubernetes.io/projected/f83ce425-101d-4489-94a4-5c256eb29328-kube-api-access-8pw52\") pod \"machine-config-daemon-7z66r\" (UID: \"f83ce425-101d-4489-94a4-5c256eb29328\") " pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.341883 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5716183c-6301-4312-923c-e34254575a82-system-cni-dir\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.341907 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-multus-socket-dir-parent\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.341957 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-multus-cni-dir\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.341981 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-multus-socket-dir-parent\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.341985 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqqm5\" (UniqueName: \"kubernetes.io/projected/5716183c-6301-4312-923c-e34254575a82-kube-api-access-qqqm5\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342104 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5716183c-6301-4312-923c-e34254575a82-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342166 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-os-release\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342200 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-host-run-multus-certs\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342267 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f83ce425-101d-4489-94a4-5c256eb29328-rootfs\") pod \"machine-config-daemon-7z66r\" (UID: \"f83ce425-101d-4489-94a4-5c256eb29328\") " pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342302 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-cnibin\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342353 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-host-run-multus-certs\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342358 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-host-var-lib-cni-bin\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342407 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5716183c-6301-4312-923c-e34254575a82-cni-binary-copy\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342443 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmcjn\" (UniqueName: \"kubernetes.io/projected/8871f396-b980-46a8-9477-b3be4889638f-kube-api-access-cmcjn\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342480 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-host-run-netns\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342511 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8871f396-b980-46a8-9477-b3be4889638f-multus-daemon-config\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342529 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-cnibin\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342560 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-hostroot\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342528 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f83ce425-101d-4489-94a4-5c256eb29328-rootfs\") pod \"machine-config-daemon-7z66r\" (UID: \"f83ce425-101d-4489-94a4-5c256eb29328\") " pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342599 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-host-run-netns\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342594 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f83ce425-101d-4489-94a4-5c256eb29328-mcd-auth-proxy-config\") pod \"machine-config-daemon-7z66r\" (UID: \"f83ce425-101d-4489-94a4-5c256eb29328\") " pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342598 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-host-var-lib-cni-multus\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342675 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-host-var-lib-cni-bin\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342695 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-multus-conf-dir\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342694 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-hostroot\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.341810 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-etc-kubernetes\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342645 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-host-var-lib-cni-multus\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342738 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5716183c-6301-4312-923c-e34254575a82-cnibin\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342772 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-system-cni-dir\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342805 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-os-release\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342766 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-host-run-k8s-cni-cncf-io\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342801 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5716183c-6301-4312-923c-e34254575a82-os-release\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342881 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5716183c-6301-4312-923c-e34254575a82-os-release\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342886 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-host-var-lib-kubelet\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342909 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8871f396-b980-46a8-9477-b3be4889638f-host-var-lib-kubelet\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.342937 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5716183c-6301-4312-923c-e34254575a82-cnibin\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.343453 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5716183c-6301-4312-923c-e34254575a82-tuning-conf-dir\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.343636 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5716183c-6301-4312-923c-e34254575a82-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.343645 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8871f396-b980-46a8-9477-b3be4889638f-multus-daemon-config\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.343804 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5716183c-6301-4312-923c-e34254575a82-cni-binary-copy\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.343832 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8871f396-b980-46a8-9477-b3be4889638f-cni-binary-copy\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.345835 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f83ce425-101d-4489-94a4-5c256eb29328-proxy-tls\") pod \"machine-config-daemon-7z66r\" (UID: \"f83ce425-101d-4489-94a4-5c256eb29328\") " pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.358255 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.361213 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pw52\" (UniqueName: \"kubernetes.io/projected/f83ce425-101d-4489-94a4-5c256eb29328-kube-api-access-8pw52\") pod \"machine-config-daemon-7z66r\" (UID: \"f83ce425-101d-4489-94a4-5c256eb29328\") " pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.370266 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqqm5\" (UniqueName: \"kubernetes.io/projected/5716183c-6301-4312-923c-e34254575a82-kube-api-access-qqqm5\") pod \"multus-additional-cni-plugins-75l8m\" (UID: \"5716183c-6301-4312-923c-e34254575a82\") " pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.372789 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmcjn\" (UniqueName: \"kubernetes.io/projected/8871f396-b980-46a8-9477-b3be4889638f-kube-api-access-cmcjn\") pod \"multus-7d9g4\" (UID: \"8871f396-b980-46a8-9477-b3be4889638f\") " pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.373260 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.390707 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.405790 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-7d9g4" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.411674 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-75l8m" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.416190 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.417295 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 16:40:55 crc kubenswrapper[4808]: W1002 16:40:55.430902 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8871f396_b980_46a8_9477_b3be4889638f.slice/crio-6f43f44f0f8ab018f30fa7348c8254e946b4f1c3d806d95f46db9a37846e0238 WatchSource:0}: Error finding container 6f43f44f0f8ab018f30fa7348c8254e946b4f1c3d806d95f46db9a37846e0238: Status 404 returned error can't find the container with id 6f43f44f0f8ab018f30fa7348c8254e946b4f1c3d806d95f46db9a37846e0238 Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.434163 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: W1002 16:40:55.440880 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5716183c_6301_4312_923c_e34254575a82.slice/crio-c9c86173d3221b75684bb78d0beceaae2cdb3b4d024c8c7f50e670ee1e307ea5 WatchSource:0}: Error finding container c9c86173d3221b75684bb78d0beceaae2cdb3b4d024c8c7f50e670ee1e307ea5: Status 404 returned error can't find the container with id c9c86173d3221b75684bb78d0beceaae2cdb3b4d024c8c7f50e670ee1e307ea5 Oct 02 16:40:55 crc kubenswrapper[4808]: W1002 16:40:55.442038 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf83ce425_101d_4489_94a4_5c256eb29328.slice/crio-b3e119dfafaf621db258966470c9621fafd385893306f2da9cb84a05bc7dff67 WatchSource:0}: Error finding container b3e119dfafaf621db258966470c9621fafd385893306f2da9cb84a05bc7dff67: Status 404 returned error can't find the container with id b3e119dfafaf621db258966470c9621fafd385893306f2da9cb84a05bc7dff67 Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.451409 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.469661 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zm46w"] Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.470672 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.472422 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.474805 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.474874 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.474946 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.475056 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.475055 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.475177 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.487620 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.500354 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.516046 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.526735 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" event={"ID":"f83ce425-101d-4489-94a4-5c256eb29328","Type":"ContainerStarted","Data":"b3e119dfafaf621db258966470c9621fafd385893306f2da9cb84a05bc7dff67"} Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.528427 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-ppf7n" event={"ID":"96534061-3b16-458d-813f-507d5d0acbcc","Type":"ContainerStarted","Data":"2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2"} Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.528479 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-ppf7n" event={"ID":"96534061-3b16-458d-813f-507d5d0acbcc","Type":"ContainerStarted","Data":"0ee456decc345f85fbe2cc212005da4d34b2499001d1ff6d720e7821820032e4"} Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.531728 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" event={"ID":"5716183c-6301-4312-923c-e34254575a82","Type":"ContainerStarted","Data":"c9c86173d3221b75684bb78d0beceaae2cdb3b4d024c8c7f50e670ee1e307ea5"} Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.532139 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.533436 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7d9g4" event={"ID":"8871f396-b980-46a8-9477-b3be4889638f","Type":"ContainerStarted","Data":"6f43f44f0f8ab018f30fa7348c8254e946b4f1c3d806d95f46db9a37846e0238"} Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.535481 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599"} Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.546446 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.569157 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.599590 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.630765 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.645693 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e441e501-09ee-4e3d-a763-c853de09a02c-ovnkube-script-lib\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.645733 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-run-openvswitch\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.645751 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-systemd-units\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.645766 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e441e501-09ee-4e3d-a763-c853de09a02c-ovnkube-config\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.645783 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-node-log\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.645796 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwkmd\" (UniqueName: \"kubernetes.io/projected/e441e501-09ee-4e3d-a763-c853de09a02c-kube-api-access-bwkmd\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.645814 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e441e501-09ee-4e3d-a763-c853de09a02c-ovn-node-metrics-cert\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.645830 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-run-ovn-kubernetes\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.645846 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e441e501-09ee-4e3d-a763-c853de09a02c-env-overrides\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.645877 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-run-ovn\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.645890 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-cni-netd\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.645911 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-kubelet\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.645934 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-etc-openvswitch\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.645947 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-cni-bin\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.645963 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-run-systemd\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.645977 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-run-netns\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.646003 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.646025 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-slash\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.646039 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-log-socket\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.646055 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-var-lib-openvswitch\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.659539 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.675794 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.688268 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.708168 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.724162 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.738801 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747213 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e441e501-09ee-4e3d-a763-c853de09a02c-ovn-node-metrics-cert\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747300 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-run-ovn-kubernetes\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747327 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e441e501-09ee-4e3d-a763-c853de09a02c-env-overrides\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747354 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-run-ovn\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747377 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-cni-netd\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747406 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-kubelet\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747445 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-cni-bin\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747469 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-run-systemd\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747476 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-cni-netd\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747490 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-etc-openvswitch\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747503 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-kubelet\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747526 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-run-systemd\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747511 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-run-netns\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747539 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-run-netns\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747540 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-etc-openvswitch\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747566 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747493 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-run-ovn-kubernetes\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747545 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-run-ovn\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747604 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-slash\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747555 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-cni-bin\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747657 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747688 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-slash\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747733 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-log-socket\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747814 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-log-socket\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747870 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-var-lib-openvswitch\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747954 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e441e501-09ee-4e3d-a763-c853de09a02c-ovnkube-script-lib\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.747979 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-run-openvswitch\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.748003 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-systemd-units\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.748041 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e441e501-09ee-4e3d-a763-c853de09a02c-ovnkube-config\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.748075 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-node-log\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.748097 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwkmd\" (UniqueName: \"kubernetes.io/projected/e441e501-09ee-4e3d-a763-c853de09a02c-kube-api-access-bwkmd\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.748167 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e441e501-09ee-4e3d-a763-c853de09a02c-env-overrides\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.748227 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-var-lib-openvswitch\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.748288 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-systemd-units\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.748318 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-node-log\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.748370 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-run-openvswitch\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.748898 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e441e501-09ee-4e3d-a763-c853de09a02c-ovnkube-config\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.748897 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e441e501-09ee-4e3d-a763-c853de09a02c-ovnkube-script-lib\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.752763 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.754687 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e441e501-09ee-4e3d-a763-c853de09a02c-ovn-node-metrics-cert\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.763624 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwkmd\" (UniqueName: \"kubernetes.io/projected/e441e501-09ee-4e3d-a763-c853de09a02c-kube-api-access-bwkmd\") pod \"ovnkube-node-zm46w\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.773212 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.786982 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.800834 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.803742 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: W1002 16:40:55.815023 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode441e501_09ee_4e3d_a763_c853de09a02c.slice/crio-810f98ce29b314aea121046f13d41ba78b4a3772059264fa2f87dad04eb64289 WatchSource:0}: Error finding container 810f98ce29b314aea121046f13d41ba78b4a3772059264fa2f87dad04eb64289: Status 404 returned error can't find the container with id 810f98ce29b314aea121046f13d41ba78b4a3772059264fa2f87dad04eb64289 Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.818872 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.833834 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.846963 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.860592 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.900841 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.940616 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.951292 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.951505 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.951580 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:40:55 crc kubenswrapper[4808]: E1002 16:40:55.951718 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 16:40:55 crc kubenswrapper[4808]: E1002 16:40:55.951785 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 16:40:59.951769461 +0000 UTC m=+27.277298461 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 16:40:55 crc kubenswrapper[4808]: E1002 16:40:55.951897 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:40:59.951888914 +0000 UTC m=+27.277417914 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:40:55 crc kubenswrapper[4808]: E1002 16:40:55.951973 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 16:40:55 crc kubenswrapper[4808]: E1002 16:40:55.952006 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 16:40:59.951999867 +0000 UTC m=+27.277528867 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 16:40:55 crc kubenswrapper[4808]: I1002 16:40:55.976655 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:55Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.018365 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.052965 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.053032 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:40:56 crc kubenswrapper[4808]: E1002 16:40:56.053193 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 16:40:56 crc kubenswrapper[4808]: E1002 16:40:56.053216 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 16:40:56 crc kubenswrapper[4808]: E1002 16:40:56.053260 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:40:56 crc kubenswrapper[4808]: E1002 16:40:56.053346 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 16:41:00.053324363 +0000 UTC m=+27.378853363 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:40:56 crc kubenswrapper[4808]: E1002 16:40:56.053759 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 16:40:56 crc kubenswrapper[4808]: E1002 16:40:56.053859 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 16:40:56 crc kubenswrapper[4808]: E1002 16:40:56.053927 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:40:56 crc kubenswrapper[4808]: E1002 16:40:56.054074 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 16:41:00.054048832 +0000 UTC m=+27.379577842 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.056161 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.102618 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.141186 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.147982 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.154315 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.171002 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.202373 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.244369 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.280258 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.320796 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.355800 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.395144 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.395248 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.395279 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:40:56 crc kubenswrapper[4808]: E1002 16:40:56.395339 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:40:56 crc kubenswrapper[4808]: E1002 16:40:56.395389 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:40:56 crc kubenswrapper[4808]: E1002 16:40:56.395431 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.398533 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.437002 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.482816 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.519652 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.545400 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7d9g4" event={"ID":"8871f396-b980-46a8-9477-b3be4889638f","Type":"ContainerStarted","Data":"ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053"} Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.548412 4808 generic.go:334] "Generic (PLEG): container finished" podID="e441e501-09ee-4e3d-a763-c853de09a02c" containerID="99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b" exitCode=0 Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.548515 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerDied","Data":"99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b"} Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.548547 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerStarted","Data":"810f98ce29b314aea121046f13d41ba78b4a3772059264fa2f87dad04eb64289"} Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.558195 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.563724 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" event={"ID":"f83ce425-101d-4489-94a4-5c256eb29328","Type":"ContainerStarted","Data":"2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e"} Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.563818 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" event={"ID":"f83ce425-101d-4489-94a4-5c256eb29328","Type":"ContainerStarted","Data":"4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba"} Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.566085 4808 generic.go:334] "Generic (PLEG): container finished" podID="5716183c-6301-4312-923c-e34254575a82" containerID="3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3" exitCode=0 Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.566191 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" event={"ID":"5716183c-6301-4312-923c-e34254575a82","Type":"ContainerDied","Data":"3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3"} Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.604159 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.634999 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.682709 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.720902 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.756746 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.800023 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.835487 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.876660 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.917581 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:56 crc kubenswrapper[4808]: I1002 16:40:56.963047 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:56Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.008314 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.060578 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.095922 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.121850 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.157736 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.203691 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.242552 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.283999 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.570151 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" event={"ID":"5716183c-6301-4312-923c-e34254575a82","Type":"ContainerStarted","Data":"d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d"} Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.573440 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerStarted","Data":"15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0"} Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.573506 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerStarted","Data":"20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6"} Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.573520 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerStarted","Data":"8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5"} Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.573531 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerStarted","Data":"964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa"} Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.573545 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerStarted","Data":"827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200"} Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.573557 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerStarted","Data":"2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f"} Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.583544 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.601795 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.616703 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.628909 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.647784 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.663615 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.685742 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.706505 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.718615 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.733631 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.746421 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.765116 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.801982 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.837256 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:57 crc kubenswrapper[4808]: I1002 16:40:57.886880 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:57Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.395652 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.395709 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:40:58 crc kubenswrapper[4808]: E1002 16:40:58.395807 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.395899 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:40:58 crc kubenswrapper[4808]: E1002 16:40:58.396146 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:40:58 crc kubenswrapper[4808]: E1002 16:40:58.396317 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.571930 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.574974 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.575048 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.575067 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.575206 4808 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.580312 4808 generic.go:334] "Generic (PLEG): container finished" podID="5716183c-6301-4312-923c-e34254575a82" containerID="d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d" exitCode=0 Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.580370 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" event={"ID":"5716183c-6301-4312-923c-e34254575a82","Type":"ContainerDied","Data":"d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d"} Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.588331 4808 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.588702 4808 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.590293 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.590397 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.590411 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.590439 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.590456 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:40:58Z","lastTransitionTime":"2025-10-02T16:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.612703 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:58Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: E1002 16:40:58.622190 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:58Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.627646 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.627721 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.627743 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.627775 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.627799 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:40:58Z","lastTransitionTime":"2025-10-02T16:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.637927 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:58Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: E1002 16:40:58.644707 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:58Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.651109 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.651154 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.651167 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.651189 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.651204 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:40:58Z","lastTransitionTime":"2025-10-02T16:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.663625 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:58Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: E1002 16:40:58.665707 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:58Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.671459 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.671501 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.671516 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.671534 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.671549 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:40:58Z","lastTransitionTime":"2025-10-02T16:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.679942 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:58Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: E1002 16:40:58.687768 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:58Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.692880 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.692920 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.692933 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.692953 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.692967 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:40:58Z","lastTransitionTime":"2025-10-02T16:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.695220 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:58Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: E1002 16:40:58.707320 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:58Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: E1002 16:40:58.707481 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.708893 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:58Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.709673 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.709739 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.709751 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.709775 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.709790 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:40:58Z","lastTransitionTime":"2025-10-02T16:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.725011 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:58Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.740917 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:58Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.756176 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:58Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.770604 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:58Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.786190 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:58Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.807124 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:58Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.811946 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.811982 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.811993 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.812009 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.812022 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:40:58Z","lastTransitionTime":"2025-10-02T16:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.823824 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:58Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.845185 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:58Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.856804 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:58Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.914922 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.914965 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.914975 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.914989 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:40:58 crc kubenswrapper[4808]: I1002 16:40:58.914999 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:40:58Z","lastTransitionTime":"2025-10-02T16:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.017986 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.018048 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.018070 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.018102 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.018125 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:40:59Z","lastTransitionTime":"2025-10-02T16:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.120783 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.120815 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.120823 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.120836 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.120845 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:40:59Z","lastTransitionTime":"2025-10-02T16:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.222744 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.222805 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.222822 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.222890 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.222909 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:40:59Z","lastTransitionTime":"2025-10-02T16:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.326050 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.326095 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.326103 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.326119 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.326129 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:40:59Z","lastTransitionTime":"2025-10-02T16:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.429030 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.429070 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.429081 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.429096 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.429106 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:40:59Z","lastTransitionTime":"2025-10-02T16:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.531371 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.531451 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.531488 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.531509 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.531524 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:40:59Z","lastTransitionTime":"2025-10-02T16:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.586979 4808 generic.go:334] "Generic (PLEG): container finished" podID="5716183c-6301-4312-923c-e34254575a82" containerID="05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57" exitCode=0 Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.587051 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" event={"ID":"5716183c-6301-4312-923c-e34254575a82","Type":"ContainerDied","Data":"05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57"} Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.612807 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:59Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.634719 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.634804 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.634831 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.634865 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.634892 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:40:59Z","lastTransitionTime":"2025-10-02T16:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.635120 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:59Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.648326 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:59Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.662624 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:59Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.674829 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:59Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.692718 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:59Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.713216 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:59Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.730259 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:59Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.738004 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.738045 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.738057 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.738079 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.738093 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:40:59Z","lastTransitionTime":"2025-10-02T16:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.751983 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:59Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.765483 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:59Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.779389 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:59Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.811507 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:59Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.831014 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:59Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.841432 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.841483 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.841501 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.841524 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.841538 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:40:59Z","lastTransitionTime":"2025-10-02T16:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.854455 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:59Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.878648 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:40:59Z is after 2025-08-24T17:21:41Z" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.944515 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.944566 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.944583 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.944606 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.944623 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:40:59Z","lastTransitionTime":"2025-10-02T16:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.996004 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.996173 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:40:59 crc kubenswrapper[4808]: I1002 16:40:59.996225 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:40:59 crc kubenswrapper[4808]: E1002 16:40:59.996411 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:41:07.996354436 +0000 UTC m=+35.321883446 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:40:59 crc kubenswrapper[4808]: E1002 16:40:59.996500 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 16:40:59 crc kubenswrapper[4808]: E1002 16:40:59.996632 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 16:41:07.996602262 +0000 UTC m=+35.322131302 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 16:40:59 crc kubenswrapper[4808]: E1002 16:40:59.996500 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 16:40:59 crc kubenswrapper[4808]: E1002 16:40:59.996761 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 16:41:07.996737275 +0000 UTC m=+35.322266265 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.047078 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.047130 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.047146 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.047169 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.047186 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:00Z","lastTransitionTime":"2025-10-02T16:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.097425 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.097500 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:00 crc kubenswrapper[4808]: E1002 16:41:00.097633 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 16:41:00 crc kubenswrapper[4808]: E1002 16:41:00.097654 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 16:41:00 crc kubenswrapper[4808]: E1002 16:41:00.097664 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:41:00 crc kubenswrapper[4808]: E1002 16:41:00.097674 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 16:41:00 crc kubenswrapper[4808]: E1002 16:41:00.097742 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 16:41:00 crc kubenswrapper[4808]: E1002 16:41:00.097717 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 16:41:08.097702192 +0000 UTC m=+35.423231192 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:41:00 crc kubenswrapper[4808]: E1002 16:41:00.097763 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:41:00 crc kubenswrapper[4808]: E1002 16:41:00.097909 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 16:41:08.097873217 +0000 UTC m=+35.423402257 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.150432 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.150478 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.150495 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.150519 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.150536 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:00Z","lastTransitionTime":"2025-10-02T16:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.253589 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.253641 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.253659 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.253681 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.253701 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:00Z","lastTransitionTime":"2025-10-02T16:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.357218 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.357751 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.357764 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.357784 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.357798 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:00Z","lastTransitionTime":"2025-10-02T16:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.395647 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.395718 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.395774 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:00 crc kubenswrapper[4808]: E1002 16:41:00.396296 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:00 crc kubenswrapper[4808]: E1002 16:41:00.396448 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:00 crc kubenswrapper[4808]: E1002 16:41:00.396570 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.460308 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.460377 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.460388 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.460409 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.460421 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:00Z","lastTransitionTime":"2025-10-02T16:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.562278 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.562343 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.562360 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.562396 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.562424 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:00Z","lastTransitionTime":"2025-10-02T16:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.594547 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerStarted","Data":"e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464"} Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.598412 4808 generic.go:334] "Generic (PLEG): container finished" podID="5716183c-6301-4312-923c-e34254575a82" containerID="7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0" exitCode=0 Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.598488 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" event={"ID":"5716183c-6301-4312-923c-e34254575a82","Type":"ContainerDied","Data":"7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0"} Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.623086 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:00Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.643114 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:00Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.656568 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:00Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.665494 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.665536 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.665548 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.665568 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.665581 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:00Z","lastTransitionTime":"2025-10-02T16:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.671617 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:00Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.686909 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:00Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.700742 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:00Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.715941 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:00Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.731068 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:00Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.744482 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:00Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.757446 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:00Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.768492 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.768518 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.768526 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.768539 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.768550 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:00Z","lastTransitionTime":"2025-10-02T16:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.776907 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:00Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.793419 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:00Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.812844 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:00Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.827542 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:00Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.849039 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:00Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.872395 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.872446 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.872458 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.872477 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.872491 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:00Z","lastTransitionTime":"2025-10-02T16:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.976094 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.976155 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.976174 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.976199 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:00 crc kubenswrapper[4808]: I1002 16:41:00.976218 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:00Z","lastTransitionTime":"2025-10-02T16:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.079730 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.079794 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.079813 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.079839 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.079858 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:01Z","lastTransitionTime":"2025-10-02T16:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.183035 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.183088 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.183104 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.183122 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.183134 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:01Z","lastTransitionTime":"2025-10-02T16:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.286979 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.287039 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.287054 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.287073 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.287091 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:01Z","lastTransitionTime":"2025-10-02T16:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.390435 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.390513 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.390531 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.390556 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.390574 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:01Z","lastTransitionTime":"2025-10-02T16:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.493447 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.493530 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.493549 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.493575 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.493595 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:01Z","lastTransitionTime":"2025-10-02T16:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.596306 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.596352 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.596364 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.596383 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.596393 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:01Z","lastTransitionTime":"2025-10-02T16:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.606092 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" event={"ID":"5716183c-6301-4312-923c-e34254575a82","Type":"ContainerStarted","Data":"437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95"} Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.631131 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:01Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.643284 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:01Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.657127 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:01Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.668548 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:01Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.683809 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:01Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.699328 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.699363 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.699373 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.699400 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.699413 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:01Z","lastTransitionTime":"2025-10-02T16:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.705618 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:01Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.722523 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:01Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.745388 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:01Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.757199 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:01Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.772590 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:01Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.787537 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:01Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.802082 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.802115 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.802123 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.802138 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.802147 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:01Z","lastTransitionTime":"2025-10-02T16:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.803060 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:01Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.817863 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:01Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.834636 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:01Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.850090 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:01Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.905214 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.905759 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.905777 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.905800 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:01 crc kubenswrapper[4808]: I1002 16:41:01.905814 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:01Z","lastTransitionTime":"2025-10-02T16:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.008411 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.008467 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.008481 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.008499 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.008512 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:02Z","lastTransitionTime":"2025-10-02T16:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.111934 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.112005 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.112026 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.112056 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.112075 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:02Z","lastTransitionTime":"2025-10-02T16:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.215456 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.215518 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.215540 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.215563 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.215580 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:02Z","lastTransitionTime":"2025-10-02T16:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.319030 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.319086 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.319098 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.319115 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.319127 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:02Z","lastTransitionTime":"2025-10-02T16:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.395450 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.395527 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:02 crc kubenswrapper[4808]: E1002 16:41:02.395591 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:02 crc kubenswrapper[4808]: E1002 16:41:02.395740 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.395450 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:02 crc kubenswrapper[4808]: E1002 16:41:02.395886 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.422097 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.422166 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.422184 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.422211 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.422274 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:02Z","lastTransitionTime":"2025-10-02T16:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.525655 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.525736 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.525754 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.525780 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.525798 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:02Z","lastTransitionTime":"2025-10-02T16:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.614910 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerStarted","Data":"0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c"} Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.615492 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.619661 4808 generic.go:334] "Generic (PLEG): container finished" podID="5716183c-6301-4312-923c-e34254575a82" containerID="437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95" exitCode=0 Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.619737 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" event={"ID":"5716183c-6301-4312-923c-e34254575a82","Type":"ContainerDied","Data":"437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95"} Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.631215 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.631270 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.631281 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.631298 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.631310 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:02Z","lastTransitionTime":"2025-10-02T16:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.637277 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.651260 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.658833 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.675270 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.689910 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.707939 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.731224 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.733527 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.733569 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.733579 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.733597 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.733606 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:02Z","lastTransitionTime":"2025-10-02T16:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.744538 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.764370 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.780539 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.793071 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.804387 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.816408 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.829890 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.835972 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.836011 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.836022 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.836039 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.836051 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:02Z","lastTransitionTime":"2025-10-02T16:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.843383 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.855799 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.871295 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.883423 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.900683 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.918700 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.930647 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.938878 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.938936 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.938947 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.938960 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.938969 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:02Z","lastTransitionTime":"2025-10-02T16:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.940846 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.957614 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.971403 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:02 crc kubenswrapper[4808]: I1002 16:41:02.993437 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:02Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.010852 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.042204 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.042261 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.042270 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.042287 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.042295 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:03Z","lastTransitionTime":"2025-10-02T16:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.042538 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.065221 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.081165 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.094030 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.108031 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.145295 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.145337 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.145349 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.145367 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.145378 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:03Z","lastTransitionTime":"2025-10-02T16:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.252017 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.252097 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.252130 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.252161 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.252185 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:03Z","lastTransitionTime":"2025-10-02T16:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.356992 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.357051 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.357070 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.357095 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.357113 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:03Z","lastTransitionTime":"2025-10-02T16:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.416103 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.438488 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.453742 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.460158 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.460317 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.460349 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.460381 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.460416 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:03Z","lastTransitionTime":"2025-10-02T16:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.470482 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.495133 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.514527 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.533796 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.556600 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.563266 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.563342 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.563359 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.563391 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.563411 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:03Z","lastTransitionTime":"2025-10-02T16:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.578904 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.599962 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.621952 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.628388 4808 generic.go:334] "Generic (PLEG): container finished" podID="5716183c-6301-4312-923c-e34254575a82" containerID="6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc" exitCode=0 Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.628464 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" event={"ID":"5716183c-6301-4312-923c-e34254575a82","Type":"ContainerDied","Data":"6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc"} Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.628582 4808 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.629033 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.642297 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.667362 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.667768 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.667795 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.667806 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.667828 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.667844 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:03Z","lastTransitionTime":"2025-10-02T16:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.685044 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.701080 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.713918 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.733772 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.749393 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.770347 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.770389 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.770402 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.770422 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.770436 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:03Z","lastTransitionTime":"2025-10-02T16:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.771833 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.788167 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.806859 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.824722 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.837657 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.846625 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.861030 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.872205 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.873719 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.873759 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.873771 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.873792 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.873805 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:03Z","lastTransitionTime":"2025-10-02T16:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.885585 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.905814 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.919771 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.939785 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.955363 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.976937 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.976999 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.977010 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.977030 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:03 crc kubenswrapper[4808]: I1002 16:41:03.977046 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:03Z","lastTransitionTime":"2025-10-02T16:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.079823 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.080252 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.080267 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.080287 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.080303 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:04Z","lastTransitionTime":"2025-10-02T16:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.182546 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.182590 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.182602 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.182618 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.182630 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:04Z","lastTransitionTime":"2025-10-02T16:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.284977 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.285010 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.285021 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.285039 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.285051 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:04Z","lastTransitionTime":"2025-10-02T16:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.388395 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.388464 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.388482 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.388506 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.388523 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:04Z","lastTransitionTime":"2025-10-02T16:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.395679 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.395705 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.395780 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:04 crc kubenswrapper[4808]: E1002 16:41:04.395911 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:04 crc kubenswrapper[4808]: E1002 16:41:04.396048 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:04 crc kubenswrapper[4808]: E1002 16:41:04.396153 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.491548 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.491606 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.491617 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.491638 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.491651 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:04Z","lastTransitionTime":"2025-10-02T16:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.594510 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.594561 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.594576 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.594594 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.594608 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:04Z","lastTransitionTime":"2025-10-02T16:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.635979 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" event={"ID":"5716183c-6301-4312-923c-e34254575a82","Type":"ContainerStarted","Data":"7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219"} Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.636059 4808 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.656487 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:04Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.673053 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:04Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.687507 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:04Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.697257 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.697301 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.697314 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.697346 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.697362 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:04Z","lastTransitionTime":"2025-10-02T16:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.705290 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:04Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.717593 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:04Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.729682 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:04Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.741737 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:04Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.759020 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:04Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.778792 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:04Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.790447 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:04Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.799867 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.799920 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.799930 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.799949 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.799962 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:04Z","lastTransitionTime":"2025-10-02T16:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.808387 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:04Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.822935 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:04Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.852905 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:04Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.871361 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:04Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.892409 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:04Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.902519 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.902580 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.902593 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.902615 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:04 crc kubenswrapper[4808]: I1002 16:41:04.902626 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:04Z","lastTransitionTime":"2025-10-02T16:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.005414 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.005480 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.005502 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.005538 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.005561 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:05Z","lastTransitionTime":"2025-10-02T16:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.109007 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.109092 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.109118 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.109150 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.109175 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:05Z","lastTransitionTime":"2025-10-02T16:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.212724 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.212798 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.212822 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.212854 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.212876 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:05Z","lastTransitionTime":"2025-10-02T16:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.316078 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.316182 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.316208 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.316281 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.316321 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:05Z","lastTransitionTime":"2025-10-02T16:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.419537 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.419579 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.419590 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.419607 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.419623 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:05Z","lastTransitionTime":"2025-10-02T16:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.522400 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.522476 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.522500 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.522531 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.522556 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:05Z","lastTransitionTime":"2025-10-02T16:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.626459 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.626525 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.626542 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.626565 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.626583 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:05Z","lastTransitionTime":"2025-10-02T16:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.643321 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zm46w_e441e501-09ee-4e3d-a763-c853de09a02c/ovnkube-controller/0.log" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.648113 4808 generic.go:334] "Generic (PLEG): container finished" podID="e441e501-09ee-4e3d-a763-c853de09a02c" containerID="0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c" exitCode=1 Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.648192 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerDied","Data":"0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c"} Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.650108 4808 scope.go:117] "RemoveContainer" containerID="0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.671497 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:05Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.693469 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:05Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.711789 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:05Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.726809 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:05Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.729164 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.729202 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.729214 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.729265 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.729303 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:05Z","lastTransitionTime":"2025-10-02T16:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.744836 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:05Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.759287 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:05Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.775082 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:05Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.804701 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:05Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.826660 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:05Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.832897 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.832935 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.832948 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.832966 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.832980 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:05Z","lastTransitionTime":"2025-10-02T16:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.853519 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"message\\\":\\\"externalversions/factory.go:141\\\\nI1002 16:41:04.669153 6078 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669327 6078 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669107 6078 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 16:41:04.669387 6078 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 16:41:04.669120 6078 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 16:41:04.669100 6078 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 16:41:04.669969 6078 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 16:41:04.670126 6078 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 16:41:04.670151 6078 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 16:41:04.670128 6078 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 16:41:04.670261 6078 factory.go:656] Stopping watch factory\\\\nI1002 16:41:04.670316 6078 ovnkube.go:599] Stopped ovnkube\\\\nI1002 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:05Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.872324 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:05Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.886860 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:05Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.902726 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:05Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.903162 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.916656 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:05Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.935049 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:05Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.935470 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.935511 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.935522 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.935548 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.935561 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:05Z","lastTransitionTime":"2025-10-02T16:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.950413 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:05Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.970960 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:05Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:05 crc kubenswrapper[4808]: I1002 16:41:05.986788 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:05Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.004003 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.025728 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.038990 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.039029 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.039042 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.039060 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.039072 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:06Z","lastTransitionTime":"2025-10-02T16:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.041698 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.069407 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.101857 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.120081 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.142137 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.142194 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.142210 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.142253 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.142273 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:06Z","lastTransitionTime":"2025-10-02T16:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.149885 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"message\\\":\\\"externalversions/factory.go:141\\\\nI1002 16:41:04.669153 6078 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669327 6078 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669107 6078 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 16:41:04.669387 6078 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 16:41:04.669120 6078 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 16:41:04.669100 6078 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 16:41:04.669969 6078 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 16:41:04.670126 6078 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 16:41:04.670151 6078 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 16:41:04.670128 6078 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 16:41:04.670261 6078 factory.go:656] Stopping watch factory\\\\nI1002 16:41:04.670316 6078 ovnkube.go:599] Stopped ovnkube\\\\nI1002 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.192604 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.206134 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.218071 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.229094 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.241426 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.245181 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.245223 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.245248 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.245267 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.245281 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:06Z","lastTransitionTime":"2025-10-02T16:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.348283 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.348607 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.348681 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.348758 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.348827 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:06Z","lastTransitionTime":"2025-10-02T16:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.396156 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.396193 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.396155 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:06 crc kubenswrapper[4808]: E1002 16:41:06.396381 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:06 crc kubenswrapper[4808]: E1002 16:41:06.396305 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:06 crc kubenswrapper[4808]: E1002 16:41:06.396657 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.451692 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.451735 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.451746 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.451762 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.451773 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:06Z","lastTransitionTime":"2025-10-02T16:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.555357 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.555695 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.555805 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.555885 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.555958 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:06Z","lastTransitionTime":"2025-10-02T16:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.654732 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zm46w_e441e501-09ee-4e3d-a763-c853de09a02c/ovnkube-controller/0.log" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.663286 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerStarted","Data":"6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109"} Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.663664 4808 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.665751 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.665809 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.665827 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.665853 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.665873 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:06Z","lastTransitionTime":"2025-10-02T16:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.690574 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.710327 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.731647 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.749590 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.770579 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.770670 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.770710 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.770759 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.770798 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:06Z","lastTransitionTime":"2025-10-02T16:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.771950 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.806853 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"message\\\":\\\"externalversions/factory.go:141\\\\nI1002 16:41:04.669153 6078 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669327 6078 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669107 6078 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 16:41:04.669387 6078 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 16:41:04.669120 6078 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 16:41:04.669100 6078 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 16:41:04.669969 6078 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 16:41:04.670126 6078 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 16:41:04.670151 6078 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 16:41:04.670128 6078 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 16:41:04.670261 6078 factory.go:656] Stopping watch factory\\\\nI1002 16:41:04.670316 6078 ovnkube.go:599] Stopped ovnkube\\\\nI1002 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.841419 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.874795 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.874857 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.874869 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.874892 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.874907 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:06Z","lastTransitionTime":"2025-10-02T16:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.875702 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.895765 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.915640 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.936832 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.952915 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.973703 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.978059 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.978385 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.978563 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.978715 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.978846 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:06Z","lastTransitionTime":"2025-10-02T16:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:06 crc kubenswrapper[4808]: I1002 16:41:06.994315 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.014185 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.081995 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.082043 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.082054 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.082072 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.082082 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:07Z","lastTransitionTime":"2025-10-02T16:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.184937 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.184995 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.185009 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.185029 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.185042 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:07Z","lastTransitionTime":"2025-10-02T16:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.274598 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq"] Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.275482 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.278795 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.278901 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.288210 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.288290 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.288312 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.288337 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.288358 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:07Z","lastTransitionTime":"2025-10-02T16:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.296346 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.313448 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.328722 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.348410 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.369867 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.376433 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4021608b-add3-4b37-b523-f1d3bf65ca05-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-llclq\" (UID: \"4021608b-add3-4b37-b523-f1d3bf65ca05\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.376711 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd76q\" (UniqueName: \"kubernetes.io/projected/4021608b-add3-4b37-b523-f1d3bf65ca05-kube-api-access-vd76q\") pod \"ovnkube-control-plane-749d76644c-llclq\" (UID: \"4021608b-add3-4b37-b523-f1d3bf65ca05\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.376861 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4021608b-add3-4b37-b523-f1d3bf65ca05-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-llclq\" (UID: \"4021608b-add3-4b37-b523-f1d3bf65ca05\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.377001 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4021608b-add3-4b37-b523-f1d3bf65ca05-env-overrides\") pod \"ovnkube-control-plane-749d76644c-llclq\" (UID: \"4021608b-add3-4b37-b523-f1d3bf65ca05\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.389813 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.391360 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.391388 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.391398 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.391418 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.391429 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:07Z","lastTransitionTime":"2025-10-02T16:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.413160 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.429646 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.442077 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.468581 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.478367 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4021608b-add3-4b37-b523-f1d3bf65ca05-env-overrides\") pod \"ovnkube-control-plane-749d76644c-llclq\" (UID: \"4021608b-add3-4b37-b523-f1d3bf65ca05\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.478540 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4021608b-add3-4b37-b523-f1d3bf65ca05-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-llclq\" (UID: \"4021608b-add3-4b37-b523-f1d3bf65ca05\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.478611 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd76q\" (UniqueName: \"kubernetes.io/projected/4021608b-add3-4b37-b523-f1d3bf65ca05-kube-api-access-vd76q\") pod \"ovnkube-control-plane-749d76644c-llclq\" (UID: \"4021608b-add3-4b37-b523-f1d3bf65ca05\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.478679 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4021608b-add3-4b37-b523-f1d3bf65ca05-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-llclq\" (UID: \"4021608b-add3-4b37-b523-f1d3bf65ca05\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.479153 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4021608b-add3-4b37-b523-f1d3bf65ca05-env-overrides\") pod \"ovnkube-control-plane-749d76644c-llclq\" (UID: \"4021608b-add3-4b37-b523-f1d3bf65ca05\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.480048 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4021608b-add3-4b37-b523-f1d3bf65ca05-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-llclq\" (UID: \"4021608b-add3-4b37-b523-f1d3bf65ca05\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.484043 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.486685 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4021608b-add3-4b37-b523-f1d3bf65ca05-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-llclq\" (UID: \"4021608b-add3-4b37-b523-f1d3bf65ca05\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.496015 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.496193 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.496280 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.496367 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.496430 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:07Z","lastTransitionTime":"2025-10-02T16:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.502315 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.503075 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd76q\" (UniqueName: \"kubernetes.io/projected/4021608b-add3-4b37-b523-f1d3bf65ca05-kube-api-access-vd76q\") pod \"ovnkube-control-plane-749d76644c-llclq\" (UID: \"4021608b-add3-4b37-b523-f1d3bf65ca05\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.528993 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.545262 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.573168 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"message\\\":\\\"externalversions/factory.go:141\\\\nI1002 16:41:04.669153 6078 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669327 6078 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669107 6078 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 16:41:04.669387 6078 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 16:41:04.669120 6078 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 16:41:04.669100 6078 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 16:41:04.669969 6078 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 16:41:04.670126 6078 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 16:41:04.670151 6078 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 16:41:04.670128 6078 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 16:41:04.670261 6078 factory.go:656] Stopping watch factory\\\\nI1002 16:41:04.670316 6078 ovnkube.go:599] Stopped ovnkube\\\\nI1002 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.590765 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.593844 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.599626 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.599716 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.599731 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.599755 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.599768 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:07Z","lastTransitionTime":"2025-10-02T16:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:07 crc kubenswrapper[4808]: W1002 16:41:07.611257 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4021608b_add3_4b37_b523_f1d3bf65ca05.slice/crio-63b8471bfa6a77bfcd794eb96e3cadcd5660f9299c6035c963b5d75180a8c146 WatchSource:0}: Error finding container 63b8471bfa6a77bfcd794eb96e3cadcd5660f9299c6035c963b5d75180a8c146: Status 404 returned error can't find the container with id 63b8471bfa6a77bfcd794eb96e3cadcd5660f9299c6035c963b5d75180a8c146 Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.668654 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" event={"ID":"4021608b-add3-4b37-b523-f1d3bf65ca05","Type":"ContainerStarted","Data":"63b8471bfa6a77bfcd794eb96e3cadcd5660f9299c6035c963b5d75180a8c146"} Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.668716 4808 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.701856 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.701897 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.701906 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.701921 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.701933 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:07Z","lastTransitionTime":"2025-10-02T16:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.805314 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.805353 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.805364 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.805380 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.805392 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:07Z","lastTransitionTime":"2025-10-02T16:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.907836 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.907872 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.907884 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.907902 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:07 crc kubenswrapper[4808]: I1002 16:41:07.907916 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:07Z","lastTransitionTime":"2025-10-02T16:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.010479 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.010516 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.010523 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.010537 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.010545 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:08Z","lastTransitionTime":"2025-10-02T16:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.083636 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.083786 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.083857 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.083971 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.083992 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:41:24.083961654 +0000 UTC m=+51.409490674 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.084026 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.084084 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 16:41:24.084062126 +0000 UTC m=+51.409591146 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.084116 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 16:41:24.084095027 +0000 UTC m=+51.409624087 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.113164 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.113219 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.113262 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.113289 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.113307 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:08Z","lastTransitionTime":"2025-10-02T16:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.184780 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.184889 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.185040 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.185087 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.185103 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.185104 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.185134 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.185155 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.185179 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 16:41:24.185158237 +0000 UTC m=+51.510687247 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.185225 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 16:41:24.185202738 +0000 UTC m=+51.510731778 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.216371 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.216411 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.216422 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.216441 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.216457 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:08Z","lastTransitionTime":"2025-10-02T16:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.319725 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.319796 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.319816 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.319842 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.319861 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:08Z","lastTransitionTime":"2025-10-02T16:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.395214 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.395406 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.395425 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.395568 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.395725 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.395888 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.418146 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-2x552"] Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.418752 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.418837 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.423063 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.423094 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.423103 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.423115 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.423125 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:08Z","lastTransitionTime":"2025-10-02T16:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.443625 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.464081 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.485860 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.489183 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qz8wh\" (UniqueName: \"kubernetes.io/projected/e5e40838-d098-402e-b99a-819ce5a5977c-kube-api-access-qz8wh\") pod \"network-metrics-daemon-2x552\" (UID: \"e5e40838-d098-402e-b99a-819ce5a5977c\") " pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.489275 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs\") pod \"network-metrics-daemon-2x552\" (UID: \"e5e40838-d098-402e-b99a-819ce5a5977c\") " pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.505606 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2x552" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e40838-d098-402e-b99a-819ce5a5977c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2x552\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.526174 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.526256 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.526268 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.526290 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.526300 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:08Z","lastTransitionTime":"2025-10-02T16:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.528944 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.546169 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.563002 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.586907 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.590008 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz8wh\" (UniqueName: \"kubernetes.io/projected/e5e40838-d098-402e-b99a-819ce5a5977c-kube-api-access-qz8wh\") pod \"network-metrics-daemon-2x552\" (UID: \"e5e40838-d098-402e-b99a-819ce5a5977c\") " pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.590185 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs\") pod \"network-metrics-daemon-2x552\" (UID: \"e5e40838-d098-402e-b99a-819ce5a5977c\") " pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.590460 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.590581 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs podName:e5e40838-d098-402e-b99a-819ce5a5977c nodeName:}" failed. No retries permitted until 2025-10-02 16:41:09.090547514 +0000 UTC m=+36.416076554 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs") pod "network-metrics-daemon-2x552" (UID: "e5e40838-d098-402e-b99a-819ce5a5977c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.602602 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.611405 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qz8wh\" (UniqueName: \"kubernetes.io/projected/e5e40838-d098-402e-b99a-819ce5a5977c-kube-api-access-qz8wh\") pod \"network-metrics-daemon-2x552\" (UID: \"e5e40838-d098-402e-b99a-819ce5a5977c\") " pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.620696 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.629300 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.629362 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.629372 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.629393 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.629409 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:08Z","lastTransitionTime":"2025-10-02T16:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.644071 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.667163 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.677350 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" event={"ID":"4021608b-add3-4b37-b523-f1d3bf65ca05","Type":"ContainerStarted","Data":"08ef5f79f90831cb32131db7ea8d2f305c3664bd9d06dd98c007b74378f1bf1c"} Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.677434 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" event={"ID":"4021608b-add3-4b37-b523-f1d3bf65ca05","Type":"ContainerStarted","Data":"6f7426ac250fc0c8a7afe7b809d6ba62dc69ae4ff0f2e595027f86958d505ee2"} Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.699293 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"message\\\":\\\"externalversions/factory.go:141\\\\nI1002 16:41:04.669153 6078 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669327 6078 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669107 6078 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 16:41:04.669387 6078 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 16:41:04.669120 6078 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 16:41:04.669100 6078 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 16:41:04.669969 6078 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 16:41:04.670126 6078 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 16:41:04.670151 6078 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 16:41:04.670128 6078 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 16:41:04.670261 6078 factory.go:656] Stopping watch factory\\\\nI1002 16:41:04.670316 6078 ovnkube.go:599] Stopped ovnkube\\\\nI1002 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.719835 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.732863 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.732922 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.732940 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.732965 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.732985 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:08Z","lastTransitionTime":"2025-10-02T16:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.739044 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.758229 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.775554 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.797346 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.816499 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.835509 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.835737 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.835792 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.835814 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.835839 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.835857 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:08Z","lastTransitionTime":"2025-10-02T16:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.853588 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2x552" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e40838-d098-402e-b99a-819ce5a5977c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2x552\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.873754 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.891093 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.920625 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.938881 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.938941 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.938960 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.938986 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.939004 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:08Z","lastTransitionTime":"2025-10-02T16:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.945588 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f7426ac250fc0c8a7afe7b809d6ba62dc69ae4ff0f2e595027f86958d505ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ef5f79f90831cb32131db7ea8d2f305c3664bd9d06dd98c007b74378f1bf1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.945810 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.945898 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.946023 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.946466 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.946534 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:08Z","lastTransitionTime":"2025-10-02T16:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.966456 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.970035 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.972387 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.972435 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.972454 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.972481 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.972499 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:08Z","lastTransitionTime":"2025-10-02T16:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:08 crc kubenswrapper[4808]: E1002 16:41:08.993171 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:08 crc kubenswrapper[4808]: I1002 16:41:08.996869 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:08.999967 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.000035 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.000058 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.000083 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.000103 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:09Z","lastTransitionTime":"2025-10-02T16:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:09 crc kubenswrapper[4808]: E1002 16:41:09.018921 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.023758 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.023805 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.023824 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.023847 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.023832 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.023860 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:09Z","lastTransitionTime":"2025-10-02T16:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:09 crc kubenswrapper[4808]: E1002 16:41:09.042663 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.047981 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.048034 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.048055 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.048083 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.048101 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:09Z","lastTransitionTime":"2025-10-02T16:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.055545 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"message\\\":\\\"externalversions/factory.go:141\\\\nI1002 16:41:04.669153 6078 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669327 6078 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669107 6078 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 16:41:04.669387 6078 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 16:41:04.669120 6078 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 16:41:04.669100 6078 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 16:41:04.669969 6078 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 16:41:04.670126 6078 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 16:41:04.670151 6078 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 16:41:04.670128 6078 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 16:41:04.670261 6078 factory.go:656] Stopping watch factory\\\\nI1002 16:41:04.670316 6078 ovnkube.go:599] Stopped ovnkube\\\\nI1002 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:09 crc kubenswrapper[4808]: E1002 16:41:09.061842 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:09 crc kubenswrapper[4808]: E1002 16:41:09.062056 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.064384 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.064420 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.064432 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.064449 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.064461 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:09Z","lastTransitionTime":"2025-10-02T16:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.074715 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.090121 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:09 crc kubenswrapper[4808]: E1002 16:41:09.097126 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 16:41:09 crc kubenswrapper[4808]: E1002 16:41:09.097258 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs podName:e5e40838-d098-402e-b99a-819ce5a5977c nodeName:}" failed. No retries permitted until 2025-10-02 16:41:10.097206515 +0000 UTC m=+37.422735525 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs") pod "network-metrics-daemon-2x552" (UID: "e5e40838-d098-402e-b99a-819ce5a5977c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.096965 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs\") pod \"network-metrics-daemon-2x552\" (UID: \"e5e40838-d098-402e-b99a-819ce5a5977c\") " pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.103843 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.119017 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.136632 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.167718 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.167778 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.167804 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.167834 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.167857 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:09Z","lastTransitionTime":"2025-10-02T16:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.271078 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.271143 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.271159 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.271189 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.271208 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:09Z","lastTransitionTime":"2025-10-02T16:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.374409 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.374532 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.374552 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.374586 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.374643 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:09Z","lastTransitionTime":"2025-10-02T16:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.478195 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.478282 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.478300 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.478324 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.478342 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:09Z","lastTransitionTime":"2025-10-02T16:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.581964 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.582022 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.582042 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.582068 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.582086 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:09Z","lastTransitionTime":"2025-10-02T16:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.684926 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.685587 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.685599 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.685624 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.685642 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:09Z","lastTransitionTime":"2025-10-02T16:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.789251 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.789311 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.789321 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.789340 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.789353 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:09Z","lastTransitionTime":"2025-10-02T16:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.892542 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.892605 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.892614 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.892631 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.892641 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:09Z","lastTransitionTime":"2025-10-02T16:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.995797 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.995873 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.995892 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.995922 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:09 crc kubenswrapper[4808]: I1002 16:41:09.995945 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:09Z","lastTransitionTime":"2025-10-02T16:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.099488 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.099582 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.099601 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.099633 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.099653 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:10Z","lastTransitionTime":"2025-10-02T16:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.109776 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs\") pod \"network-metrics-daemon-2x552\" (UID: \"e5e40838-d098-402e-b99a-819ce5a5977c\") " pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:10 crc kubenswrapper[4808]: E1002 16:41:10.109922 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 16:41:10 crc kubenswrapper[4808]: E1002 16:41:10.109993 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs podName:e5e40838-d098-402e-b99a-819ce5a5977c nodeName:}" failed. No retries permitted until 2025-10-02 16:41:12.109975054 +0000 UTC m=+39.435504054 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs") pod "network-metrics-daemon-2x552" (UID: "e5e40838-d098-402e-b99a-819ce5a5977c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.203778 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.203866 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.203886 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.203919 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.203942 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:10Z","lastTransitionTime":"2025-10-02T16:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.307517 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.307619 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.307657 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.307691 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.307729 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:10Z","lastTransitionTime":"2025-10-02T16:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.395479 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.395520 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.395579 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:10 crc kubenswrapper[4808]: E1002 16:41:10.395697 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:10 crc kubenswrapper[4808]: E1002 16:41:10.395850 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:10 crc kubenswrapper[4808]: E1002 16:41:10.396195 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.396229 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:10 crc kubenswrapper[4808]: E1002 16:41:10.396548 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.411171 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.411279 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.411309 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.411344 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.411365 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:10Z","lastTransitionTime":"2025-10-02T16:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.514158 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.514268 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.514287 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.514311 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.514331 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:10Z","lastTransitionTime":"2025-10-02T16:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.617210 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.617277 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.617290 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.617309 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.617321 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:10Z","lastTransitionTime":"2025-10-02T16:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.719917 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.719971 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.719987 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.720006 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.720019 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:10Z","lastTransitionTime":"2025-10-02T16:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.822963 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.823030 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.823050 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.823076 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.823097 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:10Z","lastTransitionTime":"2025-10-02T16:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.926492 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.926561 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.926579 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.926606 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:10 crc kubenswrapper[4808]: I1002 16:41:10.926625 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:10Z","lastTransitionTime":"2025-10-02T16:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.029568 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.029631 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.029647 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.029672 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.029689 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:11Z","lastTransitionTime":"2025-10-02T16:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.132995 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.133089 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.133125 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.133155 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.133182 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:11Z","lastTransitionTime":"2025-10-02T16:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.236467 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.236527 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.236547 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.236578 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.236599 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:11Z","lastTransitionTime":"2025-10-02T16:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.339892 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.339957 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.339979 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.340010 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.340032 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:11Z","lastTransitionTime":"2025-10-02T16:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.443058 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.443125 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.443147 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.443178 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.443202 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:11Z","lastTransitionTime":"2025-10-02T16:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.546674 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.546744 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.546761 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.546788 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.546806 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:11Z","lastTransitionTime":"2025-10-02T16:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.649792 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.649916 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.649943 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.649972 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.649997 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:11Z","lastTransitionTime":"2025-10-02T16:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.752896 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.752949 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.752967 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.752992 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.753009 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:11Z","lastTransitionTime":"2025-10-02T16:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.856353 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.856431 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.856452 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.856481 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.856498 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:11Z","lastTransitionTime":"2025-10-02T16:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.960071 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.960158 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.960182 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.960211 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:11 crc kubenswrapper[4808]: I1002 16:41:11.960271 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:11Z","lastTransitionTime":"2025-10-02T16:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.063302 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.063378 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.063396 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.063424 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.063443 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:12Z","lastTransitionTime":"2025-10-02T16:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.134848 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs\") pod \"network-metrics-daemon-2x552\" (UID: \"e5e40838-d098-402e-b99a-819ce5a5977c\") " pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:12 crc kubenswrapper[4808]: E1002 16:41:12.135112 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 16:41:12 crc kubenswrapper[4808]: E1002 16:41:12.135262 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs podName:e5e40838-d098-402e-b99a-819ce5a5977c nodeName:}" failed. No retries permitted until 2025-10-02 16:41:16.135213795 +0000 UTC m=+43.460742815 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs") pod "network-metrics-daemon-2x552" (UID: "e5e40838-d098-402e-b99a-819ce5a5977c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.166131 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.166184 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.166196 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.166216 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.166257 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:12Z","lastTransitionTime":"2025-10-02T16:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.225791 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.226104 4808 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.249432 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovnkube-controller" probeResult="failure" output="" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.269353 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovnkube-controller" probeResult="failure" output="" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.269601 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.269639 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.269652 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.269668 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.269683 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:12Z","lastTransitionTime":"2025-10-02T16:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.372349 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.372404 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.372415 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.372441 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.372454 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:12Z","lastTransitionTime":"2025-10-02T16:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.395379 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.395467 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.395570 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.395619 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:12 crc kubenswrapper[4808]: E1002 16:41:12.395645 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:12 crc kubenswrapper[4808]: E1002 16:41:12.395706 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:12 crc kubenswrapper[4808]: E1002 16:41:12.395792 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:12 crc kubenswrapper[4808]: E1002 16:41:12.395975 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.476021 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.476075 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.476085 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.476107 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.476118 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:12Z","lastTransitionTime":"2025-10-02T16:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.579609 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.579666 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.579680 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.579704 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.579720 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:12Z","lastTransitionTime":"2025-10-02T16:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.682541 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.682611 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.682629 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.682657 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.682679 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:12Z","lastTransitionTime":"2025-10-02T16:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.786490 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.786547 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.786560 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.786584 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.786601 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:12Z","lastTransitionTime":"2025-10-02T16:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.889872 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.889922 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.889935 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.889957 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.889971 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:12Z","lastTransitionTime":"2025-10-02T16:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.992801 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.992883 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.992917 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.992988 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:12 crc kubenswrapper[4808]: I1002 16:41:12.993011 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:12Z","lastTransitionTime":"2025-10-02T16:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.096513 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.096587 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.096611 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.096651 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.096681 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:13Z","lastTransitionTime":"2025-10-02T16:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.200564 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.200655 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.200706 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.200735 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.200783 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:13Z","lastTransitionTime":"2025-10-02T16:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.303975 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.304024 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.304074 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.304096 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.304106 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:13Z","lastTransitionTime":"2025-10-02T16:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.408648 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.408706 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.408716 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.408736 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.408749 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:13Z","lastTransitionTime":"2025-10-02T16:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.422336 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.454608 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.473845 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.507478 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"message\\\":\\\"externalversions/factory.go:141\\\\nI1002 16:41:04.669153 6078 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669327 6078 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669107 6078 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 16:41:04.669387 6078 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 16:41:04.669120 6078 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 16:41:04.669100 6078 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 16:41:04.669969 6078 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 16:41:04.670126 6078 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 16:41:04.670151 6078 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 16:41:04.670128 6078 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 16:41:04.670261 6078 factory.go:656] Stopping watch factory\\\\nI1002 16:41:04.670316 6078 ovnkube.go:599] Stopped ovnkube\\\\nI1002 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.511066 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.511115 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.511129 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.511152 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.511167 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:13Z","lastTransitionTime":"2025-10-02T16:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.524968 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.545492 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.559098 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.572740 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.586998 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.603456 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.618095 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.618185 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.618195 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.618254 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.618266 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:13Z","lastTransitionTime":"2025-10-02T16:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.624695 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.641533 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2x552" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e40838-d098-402e-b99a-819ce5a5977c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2x552\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.655986 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.676281 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.694439 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.723206 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.727490 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.729293 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.729321 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.729348 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.729374 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:13Z","lastTransitionTime":"2025-10-02T16:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.750483 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f7426ac250fc0c8a7afe7b809d6ba62dc69ae4ff0f2e595027f86958d505ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ef5f79f90831cb32131db7ea8d2f305c3664bd9d06dd98c007b74378f1bf1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.833196 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.833286 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.833298 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.833315 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.833325 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:13Z","lastTransitionTime":"2025-10-02T16:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.936032 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.936073 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.936083 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.936102 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:13 crc kubenswrapper[4808]: I1002 16:41:13.936114 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:13Z","lastTransitionTime":"2025-10-02T16:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.039762 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.039805 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.039813 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.039830 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.039840 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:14Z","lastTransitionTime":"2025-10-02T16:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.143328 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.143388 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.143406 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.143432 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.143452 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:14Z","lastTransitionTime":"2025-10-02T16:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.247185 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.247315 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.247336 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.247362 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.247383 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:14Z","lastTransitionTime":"2025-10-02T16:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.350944 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.351018 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.351043 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.351073 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.351095 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:14Z","lastTransitionTime":"2025-10-02T16:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.395894 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.395957 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:14 crc kubenswrapper[4808]: E1002 16:41:14.396078 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.396122 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:14 crc kubenswrapper[4808]: E1002 16:41:14.396331 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:14 crc kubenswrapper[4808]: E1002 16:41:14.396457 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.396717 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:14 crc kubenswrapper[4808]: E1002 16:41:14.396851 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.454452 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.454531 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.454552 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.454582 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.454606 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:14Z","lastTransitionTime":"2025-10-02T16:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.558494 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.558558 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.558575 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.558602 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.558620 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:14Z","lastTransitionTime":"2025-10-02T16:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.661287 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.661357 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.661375 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.661397 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.661414 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:14Z","lastTransitionTime":"2025-10-02T16:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.764613 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.764656 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.764669 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.764685 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.764697 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:14Z","lastTransitionTime":"2025-10-02T16:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.867323 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.867374 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.867390 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.867413 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.867429 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:14Z","lastTransitionTime":"2025-10-02T16:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.970426 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.970467 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.970477 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.970493 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:14 crc kubenswrapper[4808]: I1002 16:41:14.970503 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:14Z","lastTransitionTime":"2025-10-02T16:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.073991 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.074067 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.074090 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.074119 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.074142 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:15Z","lastTransitionTime":"2025-10-02T16:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.177461 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.177514 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.177533 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.177559 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.177577 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:15Z","lastTransitionTime":"2025-10-02T16:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.280553 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.280614 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.280632 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.280657 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.280675 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:15Z","lastTransitionTime":"2025-10-02T16:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.383637 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.383884 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.383895 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.383915 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.383934 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:15Z","lastTransitionTime":"2025-10-02T16:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.487443 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.487506 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.487518 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.487539 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.487552 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:15Z","lastTransitionTime":"2025-10-02T16:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.590570 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.590637 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.590657 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.590684 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.590703 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:15Z","lastTransitionTime":"2025-10-02T16:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.694242 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.694301 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.694312 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.694331 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.694345 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:15Z","lastTransitionTime":"2025-10-02T16:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.797651 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.797716 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.797737 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.797769 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.797790 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:15Z","lastTransitionTime":"2025-10-02T16:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.900995 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.901083 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.901101 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.901133 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:15 crc kubenswrapper[4808]: I1002 16:41:15.901153 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:15Z","lastTransitionTime":"2025-10-02T16:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.004837 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.004894 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.004906 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.004926 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.004939 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:16Z","lastTransitionTime":"2025-10-02T16:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.108287 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.108352 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.108372 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.108409 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.108423 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:16Z","lastTransitionTime":"2025-10-02T16:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.182791 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs\") pod \"network-metrics-daemon-2x552\" (UID: \"e5e40838-d098-402e-b99a-819ce5a5977c\") " pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:16 crc kubenswrapper[4808]: E1002 16:41:16.183083 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 16:41:16 crc kubenswrapper[4808]: E1002 16:41:16.183198 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs podName:e5e40838-d098-402e-b99a-819ce5a5977c nodeName:}" failed. No retries permitted until 2025-10-02 16:41:24.183165162 +0000 UTC m=+51.508694172 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs") pod "network-metrics-daemon-2x552" (UID: "e5e40838-d098-402e-b99a-819ce5a5977c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.212104 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.212179 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.212198 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.212229 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.212278 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:16Z","lastTransitionTime":"2025-10-02T16:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.315964 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.316043 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.316060 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.316086 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.316103 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:16Z","lastTransitionTime":"2025-10-02T16:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.395509 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.395656 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.395653 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.395712 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:16 crc kubenswrapper[4808]: E1002 16:41:16.395927 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:16 crc kubenswrapper[4808]: E1002 16:41:16.396171 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:16 crc kubenswrapper[4808]: E1002 16:41:16.396373 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:16 crc kubenswrapper[4808]: E1002 16:41:16.396628 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.419409 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.419466 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.419486 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.419509 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.419526 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:16Z","lastTransitionTime":"2025-10-02T16:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.522651 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.522723 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.522740 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.522772 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.522799 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:16Z","lastTransitionTime":"2025-10-02T16:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.625907 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.625994 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.626016 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.626043 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.626061 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:16Z","lastTransitionTime":"2025-10-02T16:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.728932 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.729017 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.729036 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.729114 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.729138 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:16Z","lastTransitionTime":"2025-10-02T16:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.833371 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.833458 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.833485 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.833516 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.833542 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:16Z","lastTransitionTime":"2025-10-02T16:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.937219 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.937344 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.937369 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.938036 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:16 crc kubenswrapper[4808]: I1002 16:41:16.938095 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:16Z","lastTransitionTime":"2025-10-02T16:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.041931 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.042064 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.042182 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.042370 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.042594 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:17Z","lastTransitionTime":"2025-10-02T16:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.147617 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.147690 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.147713 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.147748 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.147773 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:17Z","lastTransitionTime":"2025-10-02T16:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.251559 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.251615 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.251627 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.251649 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.251660 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:17Z","lastTransitionTime":"2025-10-02T16:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.354846 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.354934 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.354959 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.354992 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.355019 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:17Z","lastTransitionTime":"2025-10-02T16:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.457871 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.457931 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.457942 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.457964 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.457976 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:17Z","lastTransitionTime":"2025-10-02T16:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.560836 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.560903 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.560922 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.560950 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.560971 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:17Z","lastTransitionTime":"2025-10-02T16:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.664522 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.664616 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.664640 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.664670 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.664694 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:17Z","lastTransitionTime":"2025-10-02T16:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.768142 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.768200 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.768217 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.768261 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.768278 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:17Z","lastTransitionTime":"2025-10-02T16:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.872208 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.872338 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.872356 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.872381 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.872400 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:17Z","lastTransitionTime":"2025-10-02T16:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.976372 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.976435 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.976446 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.976468 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:17 crc kubenswrapper[4808]: I1002 16:41:17.976505 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:17Z","lastTransitionTime":"2025-10-02T16:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.080107 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.080170 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.080191 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.080220 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.080284 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:18Z","lastTransitionTime":"2025-10-02T16:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.184532 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.184595 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.184619 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.184649 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.184675 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:18Z","lastTransitionTime":"2025-10-02T16:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.287782 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.287819 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.287829 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.287842 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.287852 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:18Z","lastTransitionTime":"2025-10-02T16:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.390974 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.391019 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.391029 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.391044 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.391055 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:18Z","lastTransitionTime":"2025-10-02T16:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.395631 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.395687 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.395698 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.395664 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:18 crc kubenswrapper[4808]: E1002 16:41:18.395811 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:18 crc kubenswrapper[4808]: E1002 16:41:18.395967 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:18 crc kubenswrapper[4808]: E1002 16:41:18.396131 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:18 crc kubenswrapper[4808]: E1002 16:41:18.396409 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.492716 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.492753 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.492765 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.492782 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.492796 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:18Z","lastTransitionTime":"2025-10-02T16:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.595911 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.595967 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.595988 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.596012 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.596031 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:18Z","lastTransitionTime":"2025-10-02T16:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.700047 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.700112 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.700130 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.700154 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.700172 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:18Z","lastTransitionTime":"2025-10-02T16:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.804310 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.804367 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.804377 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.804398 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.804414 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:18Z","lastTransitionTime":"2025-10-02T16:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.907867 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.907964 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.907978 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.908002 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:18 crc kubenswrapper[4808]: I1002 16:41:18.908016 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:18Z","lastTransitionTime":"2025-10-02T16:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.011073 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.011152 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.011169 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.011199 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.011219 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:19Z","lastTransitionTime":"2025-10-02T16:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.091517 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.091572 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.091597 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.091625 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.091647 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:19Z","lastTransitionTime":"2025-10-02T16:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:19 crc kubenswrapper[4808]: E1002 16:41:19.114933 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:19Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.119886 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.119916 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.119924 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.119942 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.119953 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:19Z","lastTransitionTime":"2025-10-02T16:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:19 crc kubenswrapper[4808]: E1002 16:41:19.146487 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:19Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.153217 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.153303 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.153322 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.153351 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.153370 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:19Z","lastTransitionTime":"2025-10-02T16:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:19 crc kubenswrapper[4808]: E1002 16:41:19.176793 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:19Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.183843 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.183914 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.183932 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.183960 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.183980 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:19Z","lastTransitionTime":"2025-10-02T16:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:19 crc kubenswrapper[4808]: E1002 16:41:19.206040 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:19Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.212290 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.212363 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.212383 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.212414 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.212433 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:19Z","lastTransitionTime":"2025-10-02T16:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:19 crc kubenswrapper[4808]: E1002 16:41:19.235851 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:19Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:19 crc kubenswrapper[4808]: E1002 16:41:19.236107 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.238956 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.239017 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.239043 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.239073 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.239095 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:19Z","lastTransitionTime":"2025-10-02T16:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.343092 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.343144 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.343156 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.343178 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.343196 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:19Z","lastTransitionTime":"2025-10-02T16:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.446074 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.446145 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.446160 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.446183 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.446207 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:19Z","lastTransitionTime":"2025-10-02T16:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.549174 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.549224 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.549258 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.549275 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.549289 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:19Z","lastTransitionTime":"2025-10-02T16:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.652123 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.652175 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.652185 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.652202 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.652212 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:19Z","lastTransitionTime":"2025-10-02T16:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.755474 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.756103 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.756199 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.756335 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.756425 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:19Z","lastTransitionTime":"2025-10-02T16:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.860189 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.860247 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.860258 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.860276 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.860289 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:19Z","lastTransitionTime":"2025-10-02T16:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.962997 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.963041 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.963054 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.963074 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:19 crc kubenswrapper[4808]: I1002 16:41:19.963086 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:19Z","lastTransitionTime":"2025-10-02T16:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.065857 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.065901 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.065910 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.065928 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.065939 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:20Z","lastTransitionTime":"2025-10-02T16:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.170151 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.170215 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.170227 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.170268 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.170287 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:20Z","lastTransitionTime":"2025-10-02T16:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.274207 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.274309 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.274331 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.274357 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.274378 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:20Z","lastTransitionTime":"2025-10-02T16:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.378056 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.378133 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.378152 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.378183 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.378206 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:20Z","lastTransitionTime":"2025-10-02T16:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.395725 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.395825 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.395826 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:20 crc kubenswrapper[4808]: E1002 16:41:20.395896 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:20 crc kubenswrapper[4808]: E1002 16:41:20.396004 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.396029 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:20 crc kubenswrapper[4808]: E1002 16:41:20.396326 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:20 crc kubenswrapper[4808]: E1002 16:41:20.396386 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.482008 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.482082 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.482124 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.482269 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.482434 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:20Z","lastTransitionTime":"2025-10-02T16:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.586553 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.586627 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.586649 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.586677 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.586697 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:20Z","lastTransitionTime":"2025-10-02T16:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.690369 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.690437 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.690456 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.690528 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.690550 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:20Z","lastTransitionTime":"2025-10-02T16:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.793449 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.793509 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.793526 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.793551 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.793569 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:20Z","lastTransitionTime":"2025-10-02T16:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.896375 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.896420 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.896431 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.896447 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:20 crc kubenswrapper[4808]: I1002 16:41:20.896460 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:20Z","lastTransitionTime":"2025-10-02T16:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.000270 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.000326 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.000337 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.000360 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.000373 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:21Z","lastTransitionTime":"2025-10-02T16:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.103355 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.103411 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.103424 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.103446 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.103468 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:21Z","lastTransitionTime":"2025-10-02T16:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.206642 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.206684 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.206695 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.206715 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.206725 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:21Z","lastTransitionTime":"2025-10-02T16:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.310209 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.310283 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.310297 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.310320 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.310335 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:21Z","lastTransitionTime":"2025-10-02T16:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.412939 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.412990 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.413000 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.413019 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.413031 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:21Z","lastTransitionTime":"2025-10-02T16:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.516787 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.516846 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.516863 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.516890 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.516904 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:21Z","lastTransitionTime":"2025-10-02T16:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.620701 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.620766 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.620786 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.620813 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.620832 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:21Z","lastTransitionTime":"2025-10-02T16:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.723896 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.723965 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.723980 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.724047 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.724065 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:21Z","lastTransitionTime":"2025-10-02T16:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.827972 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.828056 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.828076 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.828102 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.828120 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:21Z","lastTransitionTime":"2025-10-02T16:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.931752 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.931808 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.931825 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.931848 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:21 crc kubenswrapper[4808]: I1002 16:41:21.931869 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:21Z","lastTransitionTime":"2025-10-02T16:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.035148 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.035546 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.035696 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.035808 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.035909 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:22Z","lastTransitionTime":"2025-10-02T16:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.142158 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.142266 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.142285 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.142310 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.142333 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:22Z","lastTransitionTime":"2025-10-02T16:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.245525 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.245616 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.245639 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.246118 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.246374 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:22Z","lastTransitionTime":"2025-10-02T16:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.350152 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.350201 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.350218 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.350267 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.350284 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:22Z","lastTransitionTime":"2025-10-02T16:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.396222 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.396369 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:22 crc kubenswrapper[4808]: E1002 16:41:22.396530 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.396828 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.396882 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:22 crc kubenswrapper[4808]: E1002 16:41:22.396973 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:22 crc kubenswrapper[4808]: E1002 16:41:22.397124 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:22 crc kubenswrapper[4808]: E1002 16:41:22.397390 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.452906 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.452989 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.453007 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.453033 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.453053 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:22Z","lastTransitionTime":"2025-10-02T16:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.556552 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.556616 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.556633 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.556659 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.556679 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:22Z","lastTransitionTime":"2025-10-02T16:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.660877 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.660959 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.660977 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.661428 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.661491 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:22Z","lastTransitionTime":"2025-10-02T16:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.764812 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.764876 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.764894 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.764917 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.764935 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:22Z","lastTransitionTime":"2025-10-02T16:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.868545 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.868610 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.868629 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.868657 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.868677 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:22Z","lastTransitionTime":"2025-10-02T16:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.972126 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.972200 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.972226 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.972306 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:22 crc kubenswrapper[4808]: I1002 16:41:22.972333 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:22Z","lastTransitionTime":"2025-10-02T16:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.076456 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.076529 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.076548 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.076605 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.076625 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:23Z","lastTransitionTime":"2025-10-02T16:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.180587 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.180665 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.180679 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.180701 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.180715 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:23Z","lastTransitionTime":"2025-10-02T16:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.284150 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.284217 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.284266 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.284303 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.284323 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:23Z","lastTransitionTime":"2025-10-02T16:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.387368 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.387485 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.387505 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.387529 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.387547 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:23Z","lastTransitionTime":"2025-10-02T16:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.421290 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:23Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.442798 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:23Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.459992 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:23Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.485457 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:23Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.491771 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.491830 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.491849 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.491877 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.491898 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:23Z","lastTransitionTime":"2025-10-02T16:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.509762 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f7426ac250fc0c8a7afe7b809d6ba62dc69ae4ff0f2e595027f86958d505ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ef5f79f90831cb32131db7ea8d2f305c3664bd9d06dd98c007b74378f1bf1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:23Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.535733 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:23Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.572042 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:23Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.586620 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:23Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.593898 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.593946 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.593964 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.593986 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.594000 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:23Z","lastTransitionTime":"2025-10-02T16:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.611987 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"message\\\":\\\"externalversions/factory.go:141\\\\nI1002 16:41:04.669153 6078 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669327 6078 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669107 6078 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 16:41:04.669387 6078 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 16:41:04.669120 6078 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 16:41:04.669100 6078 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 16:41:04.669969 6078 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 16:41:04.670126 6078 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 16:41:04.670151 6078 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 16:41:04.670128 6078 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 16:41:04.670261 6078 factory.go:656] Stopping watch factory\\\\nI1002 16:41:04.670316 6078 ovnkube.go:599] Stopped ovnkube\\\\nI1002 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:23Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.630086 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:23Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.655879 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:23Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.669510 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:23Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.683453 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:23Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.696293 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.696323 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.696333 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.696349 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.696360 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:23Z","lastTransitionTime":"2025-10-02T16:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.707318 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:23Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.721725 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:23Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.739598 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:23Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.754437 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2x552" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e40838-d098-402e-b99a-819ce5a5977c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2x552\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:23Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.798947 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.799072 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.799086 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.799110 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.799123 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:23Z","lastTransitionTime":"2025-10-02T16:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.902194 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.902272 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.902290 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.902342 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:23 crc kubenswrapper[4808]: I1002 16:41:23.902357 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:23Z","lastTransitionTime":"2025-10-02T16:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.004790 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.004859 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.004874 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.004898 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.004916 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:24Z","lastTransitionTime":"2025-10-02T16:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.088395 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:41:24 crc kubenswrapper[4808]: E1002 16:41:24.088649 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:41:56.088617302 +0000 UTC m=+83.414146322 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.088991 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.089096 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:24 crc kubenswrapper[4808]: E1002 16:41:24.089263 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 16:41:24 crc kubenswrapper[4808]: E1002 16:41:24.089352 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 16:41:24 crc kubenswrapper[4808]: E1002 16:41:24.089459 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 16:41:56.089380052 +0000 UTC m=+83.414909042 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 16:41:24 crc kubenswrapper[4808]: E1002 16:41:24.089563 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 16:41:56.089548996 +0000 UTC m=+83.415077996 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.108429 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.108489 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.108508 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.108539 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.108559 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:24Z","lastTransitionTime":"2025-10-02T16:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.190297 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.190362 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs\") pod \"network-metrics-daemon-2x552\" (UID: \"e5e40838-d098-402e-b99a-819ce5a5977c\") " pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.190394 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:24 crc kubenswrapper[4808]: E1002 16:41:24.190562 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 16:41:24 crc kubenswrapper[4808]: E1002 16:41:24.190581 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 16:41:24 crc kubenswrapper[4808]: E1002 16:41:24.190594 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:41:24 crc kubenswrapper[4808]: E1002 16:41:24.190655 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 16:41:56.190636686 +0000 UTC m=+83.516165686 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:41:24 crc kubenswrapper[4808]: E1002 16:41:24.190735 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 16:41:24 crc kubenswrapper[4808]: E1002 16:41:24.190798 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 16:41:24 crc kubenswrapper[4808]: E1002 16:41:24.190820 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:41:24 crc kubenswrapper[4808]: E1002 16:41:24.190926 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 16:41:56.190892543 +0000 UTC m=+83.516421573 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:41:24 crc kubenswrapper[4808]: E1002 16:41:24.190730 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 16:41:24 crc kubenswrapper[4808]: E1002 16:41:24.191014 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs podName:e5e40838-d098-402e-b99a-819ce5a5977c nodeName:}" failed. No retries permitted until 2025-10-02 16:41:40.190999805 +0000 UTC m=+67.516528835 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs") pod "network-metrics-daemon-2x552" (UID: "e5e40838-d098-402e-b99a-819ce5a5977c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.211488 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.211550 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.211563 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.211583 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.211597 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:24Z","lastTransitionTime":"2025-10-02T16:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.315095 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.315568 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.315745 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.315908 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.316650 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:24Z","lastTransitionTime":"2025-10-02T16:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.396040 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:24 crc kubenswrapper[4808]: E1002 16:41:24.396313 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.397123 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:24 crc kubenswrapper[4808]: E1002 16:41:24.397509 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.397288 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:24 crc kubenswrapper[4808]: E1002 16:41:24.397962 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.397574 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:24 crc kubenswrapper[4808]: E1002 16:41:24.398402 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.419512 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.419550 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.419561 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.419576 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.419585 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:24Z","lastTransitionTime":"2025-10-02T16:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.522718 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.522779 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.522814 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.522838 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.522854 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:24Z","lastTransitionTime":"2025-10-02T16:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.625906 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.626054 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.626075 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.626141 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.626170 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:24Z","lastTransitionTime":"2025-10-02T16:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.729878 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.729952 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.729972 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.729997 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.730013 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:24Z","lastTransitionTime":"2025-10-02T16:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.832641 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.832733 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.832767 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.832801 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.832823 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:24Z","lastTransitionTime":"2025-10-02T16:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.937088 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.937155 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.937178 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.937207 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:24 crc kubenswrapper[4808]: I1002 16:41:24.937225 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:24Z","lastTransitionTime":"2025-10-02T16:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.039879 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.039934 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.039952 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.039975 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.039992 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:25Z","lastTransitionTime":"2025-10-02T16:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.143322 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.143398 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.143411 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.143432 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.143445 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:25Z","lastTransitionTime":"2025-10-02T16:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.247302 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.248372 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.248523 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.248663 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.248772 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:25Z","lastTransitionTime":"2025-10-02T16:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.351916 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.352387 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.352525 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.352648 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.352756 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:25Z","lastTransitionTime":"2025-10-02T16:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.456251 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.456303 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.456312 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.456363 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.456375 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:25Z","lastTransitionTime":"2025-10-02T16:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.560456 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.560542 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.560565 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.560598 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.560624 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:25Z","lastTransitionTime":"2025-10-02T16:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.663804 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.663907 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.663929 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.663960 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.663980 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:25Z","lastTransitionTime":"2025-10-02T16:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.767268 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.767324 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.767337 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.767358 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.767369 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:25Z","lastTransitionTime":"2025-10-02T16:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.872134 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.872189 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.872206 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.872252 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.872273 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:25Z","lastTransitionTime":"2025-10-02T16:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.974535 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.974905 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.974972 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.975105 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:25 crc kubenswrapper[4808]: I1002 16:41:25.975177 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:25Z","lastTransitionTime":"2025-10-02T16:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.078942 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.078983 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.078995 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.079016 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.079032 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:26Z","lastTransitionTime":"2025-10-02T16:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.182887 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.182941 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.182956 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.182976 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.182993 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:26Z","lastTransitionTime":"2025-10-02T16:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.286046 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.286104 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.286115 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.286170 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.286191 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:26Z","lastTransitionTime":"2025-10-02T16:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.389816 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.389896 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.389911 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.389941 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.389962 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:26Z","lastTransitionTime":"2025-10-02T16:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.396209 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.396271 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.396332 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:26 crc kubenswrapper[4808]: E1002 16:41:26.396415 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:26 crc kubenswrapper[4808]: E1002 16:41:26.396599 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:26 crc kubenswrapper[4808]: E1002 16:41:26.396675 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.396763 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:26 crc kubenswrapper[4808]: E1002 16:41:26.396853 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.492752 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.492807 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.492819 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.492840 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.492853 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:26Z","lastTransitionTime":"2025-10-02T16:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.596405 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.596449 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.596459 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.596478 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.596489 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:26Z","lastTransitionTime":"2025-10-02T16:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.699527 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.699619 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.699638 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.699664 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.699682 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:26Z","lastTransitionTime":"2025-10-02T16:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.752482 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zm46w_e441e501-09ee-4e3d-a763-c853de09a02c/ovnkube-controller/1.log" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.753332 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zm46w_e441e501-09ee-4e3d-a763-c853de09a02c/ovnkube-controller/0.log" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.757402 4808 generic.go:334] "Generic (PLEG): container finished" podID="e441e501-09ee-4e3d-a763-c853de09a02c" containerID="6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109" exitCode=1 Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.757467 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerDied","Data":"6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109"} Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.757543 4808 scope.go:117] "RemoveContainer" containerID="0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.758731 4808 scope.go:117] "RemoveContainer" containerID="6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109" Oct 02 16:41:26 crc kubenswrapper[4808]: E1002 16:41:26.759194 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zm46w_openshift-ovn-kubernetes(e441e501-09ee-4e3d-a763-c853de09a02c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.779176 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.803433 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.803501 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.803513 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.803536 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.803551 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:26Z","lastTransitionTime":"2025-10-02T16:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.810311 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.827785 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.867568 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"message\\\":\\\"externalversions/factory.go:141\\\\nI1002 16:41:04.669153 6078 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669327 6078 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669107 6078 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 16:41:04.669387 6078 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 16:41:04.669120 6078 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 16:41:04.669100 6078 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 16:41:04.669969 6078 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 16:41:04.670126 6078 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 16:41:04.670151 6078 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 16:41:04.670128 6078 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 16:41:04.670261 6078 factory.go:656] Stopping watch factory\\\\nI1002 16:41:04.670316 6078 ovnkube.go:599] Stopped ovnkube\\\\nI1002 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:26Z\\\",\\\"message\\\":\\\" openshift.io/scc:anyuid]\\\\nI1002 16:41:26.411456 6224 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-image-registry/image-registry-697d97f7c8-9vv2b: failed to check if pod openshift-image-registry/image-registry-697d97f7c8-9vv2b is in primary UDN: could not find OVN pod annotation in map[imageregistry.operator.openshift.io/dependencies-checksum:sha256:f0beb9378fd30968608b370e3877d4c76f7539f11a5eebf44bee42a8b2dd7068 openshift.io/required-scc:restricted-v2 openshift.io/scc:restricted-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nI1002 16:41:26.411476 6224 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf: failed to check if pod openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf is in primary UDN: could not find OVN pod annotation in map[openshift.io/required-scc:nonroot-v2 openshift.io/scc:nonroot-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nE1002 16:41:26.489404 6224 shared_informer.go:316] \\\\\\\"Unhandled Error\\\\\\\" err=\\\\\\\"unable to sync caches for ovn-lb-controller\\\\\\\" logger=\\\\\\\"UnhandledError\\\\\\\"\\\\nI1002 16:41:26.490615 6224 ovnkube.go:599] Stopped ovnkube\\\\nI1002 16:41:26.490664 6224 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.891275 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.906354 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.906405 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.906415 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.906433 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.906446 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:26Z","lastTransitionTime":"2025-10-02T16:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.909617 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.925951 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.941739 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.964829 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.977895 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:26 crc kubenswrapper[4808]: I1002 16:41:26.994792 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.010600 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.010681 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.010788 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.010842 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.010868 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:27Z","lastTransitionTime":"2025-10-02T16:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.012650 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2x552" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e40838-d098-402e-b99a-819ce5a5977c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2x552\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.032328 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.048730 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.063852 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.083357 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.100969 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f7426ac250fc0c8a7afe7b809d6ba62dc69ae4ff0f2e595027f86958d505ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ef5f79f90831cb32131db7ea8d2f305c3664bd9d06dd98c007b74378f1bf1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.114431 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.114489 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.114507 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.114533 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.114551 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:27Z","lastTransitionTime":"2025-10-02T16:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.218541 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.218649 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.218674 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.218710 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.218737 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:27Z","lastTransitionTime":"2025-10-02T16:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.323078 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.323155 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.323174 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.323202 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.323222 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:27Z","lastTransitionTime":"2025-10-02T16:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.426597 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.426641 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.426650 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.426670 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.426681 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:27Z","lastTransitionTime":"2025-10-02T16:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.529882 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.529931 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.529952 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.529976 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.530019 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:27Z","lastTransitionTime":"2025-10-02T16:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.634005 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.634077 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.634095 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.634123 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.634143 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:27Z","lastTransitionTime":"2025-10-02T16:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.737690 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.737783 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.737806 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.737844 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.737868 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:27Z","lastTransitionTime":"2025-10-02T16:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.764627 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zm46w_e441e501-09ee-4e3d-a763-c853de09a02c/ovnkube-controller/1.log" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.841653 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.841740 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.841771 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.841796 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.841815 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:27Z","lastTransitionTime":"2025-10-02T16:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.945131 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.945185 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.945203 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.945290 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:27 crc kubenswrapper[4808]: I1002 16:41:27.945310 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:27Z","lastTransitionTime":"2025-10-02T16:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.048164 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.048271 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.048283 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.048305 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.048321 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:28Z","lastTransitionTime":"2025-10-02T16:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.151405 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.151450 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.151459 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.151476 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.151485 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:28Z","lastTransitionTime":"2025-10-02T16:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.255498 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.255583 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.255609 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.255643 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.255667 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:28Z","lastTransitionTime":"2025-10-02T16:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.359826 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.359898 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.359910 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.359935 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.359949 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:28Z","lastTransitionTime":"2025-10-02T16:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.395474 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.395600 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:28 crc kubenswrapper[4808]: E1002 16:41:28.395713 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.395746 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.395808 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:28 crc kubenswrapper[4808]: E1002 16:41:28.395960 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:28 crc kubenswrapper[4808]: E1002 16:41:28.396067 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:28 crc kubenswrapper[4808]: E1002 16:41:28.396175 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.462740 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.462804 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.462832 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.462863 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.462898 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:28Z","lastTransitionTime":"2025-10-02T16:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.566633 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.566713 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.566733 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.566757 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.566775 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:28Z","lastTransitionTime":"2025-10-02T16:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.670462 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.670530 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.670547 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.670577 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.670595 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:28Z","lastTransitionTime":"2025-10-02T16:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.772690 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.773141 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.773387 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.773549 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.773674 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:28Z","lastTransitionTime":"2025-10-02T16:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.881940 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.881994 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.882023 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.882052 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.882072 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:28Z","lastTransitionTime":"2025-10-02T16:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.985836 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.986358 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.986541 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.986732 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:28 crc kubenswrapper[4808]: I1002 16:41:28.986912 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:28Z","lastTransitionTime":"2025-10-02T16:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.089915 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.090079 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.090095 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.090116 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.090128 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:29Z","lastTransitionTime":"2025-10-02T16:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.193668 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.193735 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.193746 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.193766 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.193779 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:29Z","lastTransitionTime":"2025-10-02T16:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.244961 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.245047 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.245107 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.245167 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.245185 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:29Z","lastTransitionTime":"2025-10-02T16:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:29 crc kubenswrapper[4808]: E1002 16:41:29.262191 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.267950 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.268054 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.268075 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.268100 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.268124 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:29Z","lastTransitionTime":"2025-10-02T16:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:29 crc kubenswrapper[4808]: E1002 16:41:29.287014 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.293565 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.293626 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.293637 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.293658 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.293671 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:29Z","lastTransitionTime":"2025-10-02T16:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:29 crc kubenswrapper[4808]: E1002 16:41:29.311902 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.317501 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.317552 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.317564 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.317581 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.317597 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:29Z","lastTransitionTime":"2025-10-02T16:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:29 crc kubenswrapper[4808]: E1002 16:41:29.340265 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.345626 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.345768 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.345799 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.345945 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.345967 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:29Z","lastTransitionTime":"2025-10-02T16:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:29 crc kubenswrapper[4808]: E1002 16:41:29.369434 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:29 crc kubenswrapper[4808]: E1002 16:41:29.369676 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.372891 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.372966 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.372978 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.373008 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.373024 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:29Z","lastTransitionTime":"2025-10-02T16:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.475933 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.476324 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.476446 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.476558 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.476650 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:29Z","lastTransitionTime":"2025-10-02T16:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.580218 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.580350 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.580372 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.580400 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.580431 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:29Z","lastTransitionTime":"2025-10-02T16:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.683002 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.683403 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.683485 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.683555 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.683627 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:29Z","lastTransitionTime":"2025-10-02T16:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.786273 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.786315 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.786326 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.786343 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.786354 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:29Z","lastTransitionTime":"2025-10-02T16:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.889970 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.890038 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.890055 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.890090 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.890114 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:29Z","lastTransitionTime":"2025-10-02T16:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.968208 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.979146 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.987010 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.992934 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.992971 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.992983 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.993002 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:29 crc kubenswrapper[4808]: I1002 16:41:29.993015 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:29Z","lastTransitionTime":"2025-10-02T16:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.003342 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.017329 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.029179 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.047535 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.063807 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.079936 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.093268 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2x552" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e40838-d098-402e-b99a-819ce5a5977c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2x552\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.095637 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.095710 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.095726 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.095751 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.095766 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:30Z","lastTransitionTime":"2025-10-02T16:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.108518 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.123075 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.139095 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.159194 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.172702 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f7426ac250fc0c8a7afe7b809d6ba62dc69ae4ff0f2e595027f86958d505ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ef5f79f90831cb32131db7ea8d2f305c3664bd9d06dd98c007b74378f1bf1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.188742 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.198560 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.198613 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.198629 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.198651 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.198667 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:30Z","lastTransitionTime":"2025-10-02T16:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.214889 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.229843 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.262627 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"message\\\":\\\"externalversions/factory.go:141\\\\nI1002 16:41:04.669153 6078 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669327 6078 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669107 6078 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 16:41:04.669387 6078 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 16:41:04.669120 6078 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 16:41:04.669100 6078 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 16:41:04.669969 6078 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 16:41:04.670126 6078 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 16:41:04.670151 6078 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 16:41:04.670128 6078 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 16:41:04.670261 6078 factory.go:656] Stopping watch factory\\\\nI1002 16:41:04.670316 6078 ovnkube.go:599] Stopped ovnkube\\\\nI1002 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:26Z\\\",\\\"message\\\":\\\" openshift.io/scc:anyuid]\\\\nI1002 16:41:26.411456 6224 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-image-registry/image-registry-697d97f7c8-9vv2b: failed to check if pod openshift-image-registry/image-registry-697d97f7c8-9vv2b is in primary UDN: could not find OVN pod annotation in map[imageregistry.operator.openshift.io/dependencies-checksum:sha256:f0beb9378fd30968608b370e3877d4c76f7539f11a5eebf44bee42a8b2dd7068 openshift.io/required-scc:restricted-v2 openshift.io/scc:restricted-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nI1002 16:41:26.411476 6224 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf: failed to check if pod openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf is in primary UDN: could not find OVN pod annotation in map[openshift.io/required-scc:nonroot-v2 openshift.io/scc:nonroot-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nE1002 16:41:26.489404 6224 shared_informer.go:316] \\\\\\\"Unhandled Error\\\\\\\" err=\\\\\\\"unable to sync caches for ovn-lb-controller\\\\\\\" logger=\\\\\\\"UnhandledError\\\\\\\"\\\\nI1002 16:41:26.490615 6224 ovnkube.go:599] Stopped ovnkube\\\\nI1002 16:41:26.490664 6224 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.302322 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.302581 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.302700 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.302850 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.302955 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:30Z","lastTransitionTime":"2025-10-02T16:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.395569 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:30 crc kubenswrapper[4808]: E1002 16:41:30.396685 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.396776 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.396829 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.396757 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:30 crc kubenswrapper[4808]: E1002 16:41:30.397037 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:30 crc kubenswrapper[4808]: E1002 16:41:30.397155 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:30 crc kubenswrapper[4808]: E1002 16:41:30.397274 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.406600 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.406683 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.406701 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.406724 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.406737 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:30Z","lastTransitionTime":"2025-10-02T16:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.510288 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.510360 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.510375 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.510399 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.510415 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:30Z","lastTransitionTime":"2025-10-02T16:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.613753 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.613824 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.613844 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.613870 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.613889 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:30Z","lastTransitionTime":"2025-10-02T16:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.717486 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.717572 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.717583 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.717608 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.717626 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:30Z","lastTransitionTime":"2025-10-02T16:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.821522 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.821615 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.821639 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.821671 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.821697 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:30Z","lastTransitionTime":"2025-10-02T16:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.924648 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.924704 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.924715 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.924733 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:30 crc kubenswrapper[4808]: I1002 16:41:30.924747 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:30Z","lastTransitionTime":"2025-10-02T16:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.027551 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.027622 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.027643 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.027677 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.027696 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:31Z","lastTransitionTime":"2025-10-02T16:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.130887 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.130940 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.130953 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.130971 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.130984 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:31Z","lastTransitionTime":"2025-10-02T16:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.234616 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.234682 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.234699 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.234728 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.234748 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:31Z","lastTransitionTime":"2025-10-02T16:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.338937 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.339007 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.339026 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.339057 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.339075 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:31Z","lastTransitionTime":"2025-10-02T16:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.443086 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.443155 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.443175 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.443201 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.443217 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:31Z","lastTransitionTime":"2025-10-02T16:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.547371 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.547680 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.547704 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.547729 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.547743 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:31Z","lastTransitionTime":"2025-10-02T16:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.651116 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.651547 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.651706 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.651865 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.652022 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:31Z","lastTransitionTime":"2025-10-02T16:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.754869 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.754955 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.754972 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.754994 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.755017 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:31Z","lastTransitionTime":"2025-10-02T16:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.858652 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.859092 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.859381 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.859750 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.860069 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:31Z","lastTransitionTime":"2025-10-02T16:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.965636 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.965695 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.965709 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.965744 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:31 crc kubenswrapper[4808]: I1002 16:41:31.965759 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:31Z","lastTransitionTime":"2025-10-02T16:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.069886 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.069949 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.069963 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.069983 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.070006 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:32Z","lastTransitionTime":"2025-10-02T16:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.172839 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.172905 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.172916 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.172933 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.172945 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:32Z","lastTransitionTime":"2025-10-02T16:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.276477 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.276518 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.276529 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.276551 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.276564 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:32Z","lastTransitionTime":"2025-10-02T16:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.379301 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.379400 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.379422 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.379451 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.379471 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:32Z","lastTransitionTime":"2025-10-02T16:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.396045 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.396100 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.396219 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:32 crc kubenswrapper[4808]: E1002 16:41:32.396675 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.396902 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:32 crc kubenswrapper[4808]: E1002 16:41:32.397050 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:32 crc kubenswrapper[4808]: E1002 16:41:32.397283 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:32 crc kubenswrapper[4808]: E1002 16:41:32.397398 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.482842 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.482903 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.482918 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.482939 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.482952 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:32Z","lastTransitionTime":"2025-10-02T16:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.586331 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.586386 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.586405 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.586428 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.586445 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:32Z","lastTransitionTime":"2025-10-02T16:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.689763 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.689813 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.689892 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.689916 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.689933 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:32Z","lastTransitionTime":"2025-10-02T16:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.792620 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.792688 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.792706 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.792734 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.792754 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:32Z","lastTransitionTime":"2025-10-02T16:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.895718 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.895808 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.895832 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.895865 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.895884 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:32Z","lastTransitionTime":"2025-10-02T16:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.998994 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.999076 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.999101 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.999131 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:32 crc kubenswrapper[4808]: I1002 16:41:32.999154 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:32Z","lastTransitionTime":"2025-10-02T16:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.102373 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.102441 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.102460 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.102485 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.102504 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:33Z","lastTransitionTime":"2025-10-02T16:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.205318 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.205386 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.205404 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.205429 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.205448 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:33Z","lastTransitionTime":"2025-10-02T16:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.307562 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.308032 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.308112 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.308216 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.308346 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:33Z","lastTransitionTime":"2025-10-02T16:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.412757 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.412817 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.412838 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.412871 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.412889 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:33Z","lastTransitionTime":"2025-10-02T16:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.418614 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.441531 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.455737 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.478511 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.498781 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f7426ac250fc0c8a7afe7b809d6ba62dc69ae4ff0f2e595027f86958d505ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ef5f79f90831cb32131db7ea8d2f305c3664bd9d06dd98c007b74378f1bf1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.515112 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.515168 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.515178 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.515195 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.515205 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:33Z","lastTransitionTime":"2025-10-02T16:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.522000 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.555886 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.578042 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.602849 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f9569b6a4782593d2f3d1452454e4a4ad7f1a276151565362baa357b3a3518c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"message\\\":\\\"externalversions/factory.go:141\\\\nI1002 16:41:04.669153 6078 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669327 6078 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 16:41:04.669107 6078 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 16:41:04.669387 6078 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 16:41:04.669120 6078 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 16:41:04.669100 6078 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 16:41:04.669969 6078 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 16:41:04.670126 6078 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 16:41:04.670151 6078 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 16:41:04.670128 6078 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 16:41:04.670261 6078 factory.go:656] Stopping watch factory\\\\nI1002 16:41:04.670316 6078 ovnkube.go:599] Stopped ovnkube\\\\nI1002 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:26Z\\\",\\\"message\\\":\\\" openshift.io/scc:anyuid]\\\\nI1002 16:41:26.411456 6224 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-image-registry/image-registry-697d97f7c8-9vv2b: failed to check if pod openshift-image-registry/image-registry-697d97f7c8-9vv2b is in primary UDN: could not find OVN pod annotation in map[imageregistry.operator.openshift.io/dependencies-checksum:sha256:f0beb9378fd30968608b370e3877d4c76f7539f11a5eebf44bee42a8b2dd7068 openshift.io/required-scc:restricted-v2 openshift.io/scc:restricted-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nI1002 16:41:26.411476 6224 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf: failed to check if pod openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf is in primary UDN: could not find OVN pod annotation in map[openshift.io/required-scc:nonroot-v2 openshift.io/scc:nonroot-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nE1002 16:41:26.489404 6224 shared_informer.go:316] \\\\\\\"Unhandled Error\\\\\\\" err=\\\\\\\"unable to sync caches for ovn-lb-controller\\\\\\\" logger=\\\\\\\"UnhandledError\\\\\\\"\\\\nI1002 16:41:26.490615 6224 ovnkube.go:599] Stopped ovnkube\\\\nI1002 16:41:26.490664 6224 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.619122 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.619362 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.619437 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.619508 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.619573 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:33Z","lastTransitionTime":"2025-10-02T16:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.624538 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.648095 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.664435 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.681310 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.701354 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"560139b3-154f-45d4-ae7d-db62351531c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07b9557dcbd29f15e9ae34cc712160a7f5f34330deaa1c08774efd54e68c4ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28b9ee75280e7e9c44e99c1b82c9760990628488cb359247df6483530fa09ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc88fc50e88cd95a703f74122351a86eee2dc44b402e48893690ecd79c9e5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.722572 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.722668 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.722701 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.722738 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.722765 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:33Z","lastTransitionTime":"2025-10-02T16:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.722771 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.740289 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.765213 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.782637 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2x552" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e40838-d098-402e-b99a-819ce5a5977c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2x552\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.825405 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.825465 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.825479 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.825498 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.825509 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:33Z","lastTransitionTime":"2025-10-02T16:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.929050 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.929150 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.929168 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.929195 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:33 crc kubenswrapper[4808]: I1002 16:41:33.929215 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:33Z","lastTransitionTime":"2025-10-02T16:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.033127 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.033172 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.033185 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.033204 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.033216 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:34Z","lastTransitionTime":"2025-10-02T16:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.136515 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.136588 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.136600 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.136621 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.136635 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:34Z","lastTransitionTime":"2025-10-02T16:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.239666 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.239723 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.239733 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.239755 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.239766 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:34Z","lastTransitionTime":"2025-10-02T16:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.343892 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.343969 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.343987 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.344013 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.344032 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:34Z","lastTransitionTime":"2025-10-02T16:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.395785 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.395861 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.395931 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.395861 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:34 crc kubenswrapper[4808]: E1002 16:41:34.396069 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:34 crc kubenswrapper[4808]: E1002 16:41:34.396215 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:34 crc kubenswrapper[4808]: E1002 16:41:34.396324 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:34 crc kubenswrapper[4808]: E1002 16:41:34.396383 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.447759 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.447816 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.447829 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.447849 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.447861 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:34Z","lastTransitionTime":"2025-10-02T16:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.551078 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.551164 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.551182 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.551212 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.551266 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:34Z","lastTransitionTime":"2025-10-02T16:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.655808 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.655878 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.655891 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.655917 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.655930 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:34Z","lastTransitionTime":"2025-10-02T16:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.759296 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.759364 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.759378 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.759411 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.759426 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:34Z","lastTransitionTime":"2025-10-02T16:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.862786 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.862844 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.862854 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.862874 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.862884 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:34Z","lastTransitionTime":"2025-10-02T16:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.966103 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.966178 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.966196 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.966222 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:34 crc kubenswrapper[4808]: I1002 16:41:34.966311 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:34Z","lastTransitionTime":"2025-10-02T16:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.070425 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.070462 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.070474 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.070494 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.070509 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:35Z","lastTransitionTime":"2025-10-02T16:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.173712 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.173790 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.173853 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.173879 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.173892 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:35Z","lastTransitionTime":"2025-10-02T16:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.277540 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.277601 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.277610 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.277628 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.277637 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:35Z","lastTransitionTime":"2025-10-02T16:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.380864 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.380944 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.380958 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.380984 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.381000 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:35Z","lastTransitionTime":"2025-10-02T16:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.485878 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.485973 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.485993 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.486063 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.486084 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:35Z","lastTransitionTime":"2025-10-02T16:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.590802 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.590910 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.590932 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.590996 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.591014 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:35Z","lastTransitionTime":"2025-10-02T16:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.694325 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.694428 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.694452 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.694484 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.694504 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:35Z","lastTransitionTime":"2025-10-02T16:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.798390 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.798476 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.798495 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.798521 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.798553 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:35Z","lastTransitionTime":"2025-10-02T16:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.901593 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.901671 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.901685 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.901706 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:35 crc kubenswrapper[4808]: I1002 16:41:35.902006 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:35Z","lastTransitionTime":"2025-10-02T16:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.004808 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.004867 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.004882 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.004904 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.004919 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:36Z","lastTransitionTime":"2025-10-02T16:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.113764 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.113815 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.113828 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.113847 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.113860 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:36Z","lastTransitionTime":"2025-10-02T16:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.216682 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.216735 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.216748 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.216770 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.216784 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:36Z","lastTransitionTime":"2025-10-02T16:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.319368 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.319448 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.319473 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.319505 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.319527 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:36Z","lastTransitionTime":"2025-10-02T16:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.396190 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:36 crc kubenswrapper[4808]: E1002 16:41:36.396430 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.396768 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:36 crc kubenswrapper[4808]: E1002 16:41:36.396872 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.397397 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:36 crc kubenswrapper[4808]: E1002 16:41:36.397491 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.397558 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:36 crc kubenswrapper[4808]: E1002 16:41:36.397636 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.422127 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.422171 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.422189 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.422211 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.422262 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:36Z","lastTransitionTime":"2025-10-02T16:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.524988 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.525046 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.525066 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.525093 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.525109 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:36Z","lastTransitionTime":"2025-10-02T16:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.629071 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.629124 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.629135 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.629153 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.629167 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:36Z","lastTransitionTime":"2025-10-02T16:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.732594 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.732662 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.732681 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.732705 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.732720 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:36Z","lastTransitionTime":"2025-10-02T16:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.835113 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.835167 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.835179 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.835198 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.835211 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:36Z","lastTransitionTime":"2025-10-02T16:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.938380 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.938440 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.938455 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.938481 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:36 crc kubenswrapper[4808]: I1002 16:41:36.938498 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:36Z","lastTransitionTime":"2025-10-02T16:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.041549 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.041610 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.041626 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.041645 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.041657 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:37Z","lastTransitionTime":"2025-10-02T16:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.144739 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.144788 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.144801 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.144823 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.144836 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:37Z","lastTransitionTime":"2025-10-02T16:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.248053 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.248121 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.248135 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.248173 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.248188 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:37Z","lastTransitionTime":"2025-10-02T16:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.352758 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.352831 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.352858 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.352881 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.352893 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:37Z","lastTransitionTime":"2025-10-02T16:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.455266 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.455338 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.455351 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.455376 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.455392 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:37Z","lastTransitionTime":"2025-10-02T16:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.558315 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.558382 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.558392 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.558432 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.558444 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:37Z","lastTransitionTime":"2025-10-02T16:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.661719 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.661786 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.661796 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.661817 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.661835 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:37Z","lastTransitionTime":"2025-10-02T16:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.764742 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.764778 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.764787 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.764805 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.764814 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:37Z","lastTransitionTime":"2025-10-02T16:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.867164 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.867252 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.867268 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.867296 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.867311 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:37Z","lastTransitionTime":"2025-10-02T16:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.970740 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.970801 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.970810 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.970829 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:37 crc kubenswrapper[4808]: I1002 16:41:37.970843 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:37Z","lastTransitionTime":"2025-10-02T16:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.074150 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.074511 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.074796 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.074977 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.075059 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:38Z","lastTransitionTime":"2025-10-02T16:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.178296 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.178340 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.178351 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.178373 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.178387 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:38Z","lastTransitionTime":"2025-10-02T16:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.280981 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.281034 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.281046 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.281064 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.281073 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:38Z","lastTransitionTime":"2025-10-02T16:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.384747 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.384786 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.384796 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.384812 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.384824 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:38Z","lastTransitionTime":"2025-10-02T16:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.396181 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.396194 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.396214 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:38 crc kubenswrapper[4808]: E1002 16:41:38.396384 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:38 crc kubenswrapper[4808]: E1002 16:41:38.396834 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:38 crc kubenswrapper[4808]: E1002 16:41:38.396919 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.397134 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:38 crc kubenswrapper[4808]: E1002 16:41:38.397277 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.487976 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.488084 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.488105 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.488133 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.488152 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:38Z","lastTransitionTime":"2025-10-02T16:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.592128 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.592601 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.592697 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.592813 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.592922 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:38Z","lastTransitionTime":"2025-10-02T16:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.696533 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.696652 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.696680 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.696716 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.696739 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:38Z","lastTransitionTime":"2025-10-02T16:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.800086 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.800179 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.800204 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.800268 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.800295 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:38Z","lastTransitionTime":"2025-10-02T16:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.903402 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.903795 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.903871 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.903943 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:38 crc kubenswrapper[4808]: I1002 16:41:38.904015 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:38Z","lastTransitionTime":"2025-10-02T16:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.007179 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.007291 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.007307 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.007326 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.007338 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:39Z","lastTransitionTime":"2025-10-02T16:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.109457 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.109482 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.109491 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.109503 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.109512 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:39Z","lastTransitionTime":"2025-10-02T16:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.211454 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.211505 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.211519 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.211542 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.211556 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:39Z","lastTransitionTime":"2025-10-02T16:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.314827 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.314881 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.314891 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.314912 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.314925 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:39Z","lastTransitionTime":"2025-10-02T16:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.418081 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.418127 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.418138 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.418156 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.418167 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:39Z","lastTransitionTime":"2025-10-02T16:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.522077 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.522160 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.522185 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.522223 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.522283 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:39Z","lastTransitionTime":"2025-10-02T16:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.626040 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.626116 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.626134 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.626163 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.626182 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:39Z","lastTransitionTime":"2025-10-02T16:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.703965 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.704031 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.704051 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.704077 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.704097 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:39Z","lastTransitionTime":"2025-10-02T16:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:39 crc kubenswrapper[4808]: E1002 16:41:39.721172 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.726791 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.726877 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.726894 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.726929 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.726949 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:39Z","lastTransitionTime":"2025-10-02T16:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:39 crc kubenswrapper[4808]: E1002 16:41:39.746910 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.753414 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.753499 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.753528 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.753562 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.753588 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:39Z","lastTransitionTime":"2025-10-02T16:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:39 crc kubenswrapper[4808]: E1002 16:41:39.800000 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.808255 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.808642 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.808768 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.808898 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.808993 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:39Z","lastTransitionTime":"2025-10-02T16:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:39 crc kubenswrapper[4808]: E1002 16:41:39.826368 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.831113 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.831173 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.831189 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.831214 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.831248 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:39Z","lastTransitionTime":"2025-10-02T16:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:39 crc kubenswrapper[4808]: E1002 16:41:39.852985 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:39 crc kubenswrapper[4808]: E1002 16:41:39.853101 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.855350 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.855386 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.855398 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.855413 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.855424 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:39Z","lastTransitionTime":"2025-10-02T16:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.958468 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.958514 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.958524 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.958543 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:39 crc kubenswrapper[4808]: I1002 16:41:39.958555 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:39Z","lastTransitionTime":"2025-10-02T16:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.061622 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.061689 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.061705 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.061729 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.061745 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:40Z","lastTransitionTime":"2025-10-02T16:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.165904 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.166014 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.166036 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.166063 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.166080 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:40Z","lastTransitionTime":"2025-10-02T16:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.269178 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.269526 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.269645 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.269739 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.269822 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:40Z","lastTransitionTime":"2025-10-02T16:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.285908 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs\") pod \"network-metrics-daemon-2x552\" (UID: \"e5e40838-d098-402e-b99a-819ce5a5977c\") " pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:40 crc kubenswrapper[4808]: E1002 16:41:40.286123 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 16:41:40 crc kubenswrapper[4808]: E1002 16:41:40.286195 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs podName:e5e40838-d098-402e-b99a-819ce5a5977c nodeName:}" failed. No retries permitted until 2025-10-02 16:42:12.286171791 +0000 UTC m=+99.611700791 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs") pod "network-metrics-daemon-2x552" (UID: "e5e40838-d098-402e-b99a-819ce5a5977c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.372624 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.372695 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.372719 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.372748 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.372768 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:40Z","lastTransitionTime":"2025-10-02T16:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.395602 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.395725 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.395777 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:40 crc kubenswrapper[4808]: E1002 16:41:40.395840 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.395862 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:40 crc kubenswrapper[4808]: E1002 16:41:40.395918 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:40 crc kubenswrapper[4808]: E1002 16:41:40.395955 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:40 crc kubenswrapper[4808]: E1002 16:41:40.396015 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.396825 4808 scope.go:117] "RemoveContainer" containerID="6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.422317 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.448207 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.467602 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.475199 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.475265 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.475282 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.475303 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.475315 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:40Z","lastTransitionTime":"2025-10-02T16:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.492708 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:26Z\\\",\\\"message\\\":\\\" openshift.io/scc:anyuid]\\\\nI1002 16:41:26.411456 6224 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-image-registry/image-registry-697d97f7c8-9vv2b: failed to check if pod openshift-image-registry/image-registry-697d97f7c8-9vv2b is in primary UDN: could not find OVN pod annotation in map[imageregistry.operator.openshift.io/dependencies-checksum:sha256:f0beb9378fd30968608b370e3877d4c76f7539f11a5eebf44bee42a8b2dd7068 openshift.io/required-scc:restricted-v2 openshift.io/scc:restricted-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nI1002 16:41:26.411476 6224 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf: failed to check if pod openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf is in primary UDN: could not find OVN pod annotation in map[openshift.io/required-scc:nonroot-v2 openshift.io/scc:nonroot-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nE1002 16:41:26.489404 6224 shared_informer.go:316] \\\\\\\"Unhandled Error\\\\\\\" err=\\\\\\\"unable to sync caches for ovn-lb-controller\\\\\\\" logger=\\\\\\\"UnhandledError\\\\\\\"\\\\nI1002 16:41:26.490615 6224 ovnkube.go:599] Stopped ovnkube\\\\nI1002 16:41:26.490664 6224 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zm46w_openshift-ovn-kubernetes(e441e501-09ee-4e3d-a763-c853de09a02c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.508605 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.524379 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.535811 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.550166 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.569585 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"560139b3-154f-45d4-ae7d-db62351531c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07b9557dcbd29f15e9ae34cc712160a7f5f34330deaa1c08774efd54e68c4ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28b9ee75280e7e9c44e99c1b82c9760990628488cb359247df6483530fa09ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc88fc50e88cd95a703f74122351a86eee2dc44b402e48893690ecd79c9e5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.578429 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.578485 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.578506 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.578536 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.578556 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:40Z","lastTransitionTime":"2025-10-02T16:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.584499 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.595262 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.608291 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.621624 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2x552" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e40838-d098-402e-b99a-819ce5a5977c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2x552\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.639099 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.652700 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.663557 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.676912 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.680837 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.680889 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.680903 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.680944 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.680961 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:40Z","lastTransitionTime":"2025-10-02T16:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.688832 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f7426ac250fc0c8a7afe7b809d6ba62dc69ae4ff0f2e595027f86958d505ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ef5f79f90831cb32131db7ea8d2f305c3664bd9d06dd98c007b74378f1bf1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.783348 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.783402 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.783413 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.783437 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.783451 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:40Z","lastTransitionTime":"2025-10-02T16:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.825277 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zm46w_e441e501-09ee-4e3d-a763-c853de09a02c/ovnkube-controller/1.log" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.828876 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerStarted","Data":"98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8"} Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.829461 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.848768 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.872668 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.886662 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.886712 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.886722 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.886739 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.886750 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:40Z","lastTransitionTime":"2025-10-02T16:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.890975 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.911514 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.930008 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f7426ac250fc0c8a7afe7b809d6ba62dc69ae4ff0f2e595027f86958d505ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ef5f79f90831cb32131db7ea8d2f305c3664bd9d06dd98c007b74378f1bf1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.946470 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.965147 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.980080 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.988998 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.989036 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.989049 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.989066 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:40 crc kubenswrapper[4808]: I1002 16:41:40.989079 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:40Z","lastTransitionTime":"2025-10-02T16:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.002793 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:26Z\\\",\\\"message\\\":\\\" openshift.io/scc:anyuid]\\\\nI1002 16:41:26.411456 6224 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-image-registry/image-registry-697d97f7c8-9vv2b: failed to check if pod openshift-image-registry/image-registry-697d97f7c8-9vv2b is in primary UDN: could not find OVN pod annotation in map[imageregistry.operator.openshift.io/dependencies-checksum:sha256:f0beb9378fd30968608b370e3877d4c76f7539f11a5eebf44bee42a8b2dd7068 openshift.io/required-scc:restricted-v2 openshift.io/scc:restricted-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nI1002 16:41:26.411476 6224 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf: failed to check if pod openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf is in primary UDN: could not find OVN pod annotation in map[openshift.io/required-scc:nonroot-v2 openshift.io/scc:nonroot-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nE1002 16:41:26.489404 6224 shared_informer.go:316] \\\\\\\"Unhandled Error\\\\\\\" err=\\\\\\\"unable to sync caches for ovn-lb-controller\\\\\\\" logger=\\\\\\\"UnhandledError\\\\\\\"\\\\nI1002 16:41:26.490615 6224 ovnkube.go:599] Stopped ovnkube\\\\nI1002 16:41:26.490664 6224 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.019815 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.035933 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.050874 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.067212 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.079737 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"560139b3-154f-45d4-ae7d-db62351531c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07b9557dcbd29f15e9ae34cc712160a7f5f34330deaa1c08774efd54e68c4ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28b9ee75280e7e9c44e99c1b82c9760990628488cb359247df6483530fa09ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc88fc50e88cd95a703f74122351a86eee2dc44b402e48893690ecd79c9e5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.092429 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.092476 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.092490 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.092513 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.092528 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:41Z","lastTransitionTime":"2025-10-02T16:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.108641 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.130311 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.145302 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.159469 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2x552" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e40838-d098-402e-b99a-819ce5a5977c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2x552\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.195813 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.195861 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.195873 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.195893 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.195907 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:41Z","lastTransitionTime":"2025-10-02T16:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.298995 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.299052 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.299061 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.299083 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.299096 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:41Z","lastTransitionTime":"2025-10-02T16:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.401617 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.401668 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.401679 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.401695 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.401705 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:41Z","lastTransitionTime":"2025-10-02T16:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.504957 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.505018 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.505029 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.505050 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.505064 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:41Z","lastTransitionTime":"2025-10-02T16:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.608125 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.608188 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.608205 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.608250 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.608264 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:41Z","lastTransitionTime":"2025-10-02T16:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.711649 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.711712 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.711727 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.711748 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.711764 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:41Z","lastTransitionTime":"2025-10-02T16:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.814487 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.814537 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.814552 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.814571 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.814586 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:41Z","lastTransitionTime":"2025-10-02T16:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.835988 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zm46w_e441e501-09ee-4e3d-a763-c853de09a02c/ovnkube-controller/2.log" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.836806 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zm46w_e441e501-09ee-4e3d-a763-c853de09a02c/ovnkube-controller/1.log" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.840137 4808 generic.go:334] "Generic (PLEG): container finished" podID="e441e501-09ee-4e3d-a763-c853de09a02c" containerID="98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8" exitCode=1 Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.840244 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerDied","Data":"98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8"} Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.840312 4808 scope.go:117] "RemoveContainer" containerID="6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.841017 4808 scope.go:117] "RemoveContainer" containerID="98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8" Oct 02 16:41:41 crc kubenswrapper[4808]: E1002 16:41:41.841219 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zm46w_openshift-ovn-kubernetes(e441e501-09ee-4e3d-a763-c853de09a02c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.859595 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"560139b3-154f-45d4-ae7d-db62351531c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07b9557dcbd29f15e9ae34cc712160a7f5f34330deaa1c08774efd54e68c4ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28b9ee75280e7e9c44e99c1b82c9760990628488cb359247df6483530fa09ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc88fc50e88cd95a703f74122351a86eee2dc44b402e48893690ecd79c9e5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.874193 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.887201 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.902663 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.917819 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.917861 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.917875 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.917895 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.917912 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:41Z","lastTransitionTime":"2025-10-02T16:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.918296 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2x552" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e40838-d098-402e-b99a-819ce5a5977c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2x552\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.931119 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.945828 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.957888 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.973909 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:41 crc kubenswrapper[4808]: I1002 16:41:41.992801 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f7426ac250fc0c8a7afe7b809d6ba62dc69ae4ff0f2e595027f86958d505ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ef5f79f90831cb32131db7ea8d2f305c3664bd9d06dd98c007b74378f1bf1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.011678 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:42Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.022029 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.022074 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.022086 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.022106 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.022119 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:42Z","lastTransitionTime":"2025-10-02T16:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.043800 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:42Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.064171 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:42Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.084861 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:26Z\\\",\\\"message\\\":\\\" openshift.io/scc:anyuid]\\\\nI1002 16:41:26.411456 6224 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-image-registry/image-registry-697d97f7c8-9vv2b: failed to check if pod openshift-image-registry/image-registry-697d97f7c8-9vv2b is in primary UDN: could not find OVN pod annotation in map[imageregistry.operator.openshift.io/dependencies-checksum:sha256:f0beb9378fd30968608b370e3877d4c76f7539f11a5eebf44bee42a8b2dd7068 openshift.io/required-scc:restricted-v2 openshift.io/scc:restricted-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nI1002 16:41:26.411476 6224 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf: failed to check if pod openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf is in primary UDN: could not find OVN pod annotation in map[openshift.io/required-scc:nonroot-v2 openshift.io/scc:nonroot-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nE1002 16:41:26.489404 6224 shared_informer.go:316] \\\\\\\"Unhandled Error\\\\\\\" err=\\\\\\\"unable to sync caches for ovn-lb-controller\\\\\\\" logger=\\\\\\\"UnhandledError\\\\\\\"\\\\nI1002 16:41:26.490615 6224 ovnkube.go:599] Stopped ovnkube\\\\nI1002 16:41:26.490664 6224 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:41Z\\\",\\\"message\\\":\\\" failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z]\\\\nI1002 16:41:41.372050 6653 services_controller.go:434] Service openshift-machine-config-operator/machine-config-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-config-operator openshift-machine-config-operator 8bc1afc2-8724-4135-84df-aee09f23af4c 4514 0 2025-02-23 05:12:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:machine-config-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:mco-proxy-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00067dfeb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Na\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:42Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.099137 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:42Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.113595 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:42Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.124983 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.125018 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.125027 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.125041 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.125052 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:42Z","lastTransitionTime":"2025-10-02T16:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.128111 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:42Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.143226 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:42Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.227644 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.227701 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.227712 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.227734 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.227748 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:42Z","lastTransitionTime":"2025-10-02T16:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.331459 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.331512 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.331522 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.331538 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.331548 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:42Z","lastTransitionTime":"2025-10-02T16:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.395521 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:42 crc kubenswrapper[4808]: E1002 16:41:42.395705 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.395983 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:42 crc kubenswrapper[4808]: E1002 16:41:42.396098 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.396115 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.396264 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:42 crc kubenswrapper[4808]: E1002 16:41:42.396422 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:42 crc kubenswrapper[4808]: E1002 16:41:42.396559 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.435727 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.435821 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.435845 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.435870 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.435888 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:42Z","lastTransitionTime":"2025-10-02T16:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.539452 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.539512 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.539525 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.539546 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.539557 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:42Z","lastTransitionTime":"2025-10-02T16:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.642222 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.642282 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.642291 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.642309 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.642321 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:42Z","lastTransitionTime":"2025-10-02T16:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.745510 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.745565 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.745578 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.745596 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.745609 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:42Z","lastTransitionTime":"2025-10-02T16:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.845130 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7d9g4_8871f396-b980-46a8-9477-b3be4889638f/kube-multus/0.log" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.845193 4808 generic.go:334] "Generic (PLEG): container finished" podID="8871f396-b980-46a8-9477-b3be4889638f" containerID="ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053" exitCode=1 Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.845298 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7d9g4" event={"ID":"8871f396-b980-46a8-9477-b3be4889638f","Type":"ContainerDied","Data":"ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053"} Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.845757 4808 scope.go:117] "RemoveContainer" containerID="ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.847278 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.849094 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.848155 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zm46w_e441e501-09ee-4e3d-a763-c853de09a02c/ovnkube-controller/2.log" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.849120 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.849297 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.849327 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:42Z","lastTransitionTime":"2025-10-02T16:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.854750 4808 scope.go:117] "RemoveContainer" containerID="98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8" Oct 02 16:41:42 crc kubenswrapper[4808]: E1002 16:41:42.854915 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zm46w_openshift-ovn-kubernetes(e441e501-09ee-4e3d-a763-c853de09a02c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.863712 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:42Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.878289 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:42Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.890351 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:42Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.904410 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:42Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.923936 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"560139b3-154f-45d4-ae7d-db62351531c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07b9557dcbd29f15e9ae34cc712160a7f5f34330deaa1c08774efd54e68c4ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28b9ee75280e7e9c44e99c1b82c9760990628488cb359247df6483530fa09ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc88fc50e88cd95a703f74122351a86eee2dc44b402e48893690ecd79c9e5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:42Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.940574 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:42Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.952186 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.952321 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.952336 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.952365 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.952442 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:42Z","lastTransitionTime":"2025-10-02T16:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.956314 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:42Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.971054 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:42Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.985081 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2x552" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e40838-d098-402e-b99a-819ce5a5977c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2x552\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:42Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:42 crc kubenswrapper[4808]: I1002 16:41:42.997519 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:42Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.010537 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.021378 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.034887 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.047108 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f7426ac250fc0c8a7afe7b809d6ba62dc69ae4ff0f2e595027f86958d505ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ef5f79f90831cb32131db7ea8d2f305c3664bd9d06dd98c007b74378f1bf1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.055641 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.055692 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.055702 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.055717 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.055726 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:43Z","lastTransitionTime":"2025-10-02T16:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.061506 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.082836 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.095773 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:42Z\\\",\\\"message\\\":\\\"2025-10-02T16:40:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_15c55b75-b1a1-40e0-876d-f4ed62e46169\\\\n2025-10-02T16:40:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_15c55b75-b1a1-40e0-876d-f4ed62e46169 to /host/opt/cni/bin/\\\\n2025-10-02T16:40:57Z [verbose] multus-daemon started\\\\n2025-10-02T16:40:57Z [verbose] Readiness Indicator file check\\\\n2025-10-02T16:41:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.118786 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6437aef9e14da6be0715e43fc976d4402fe6b2eae51cd1673fc8ab5f96948109\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:26Z\\\",\\\"message\\\":\\\" openshift.io/scc:anyuid]\\\\nI1002 16:41:26.411456 6224 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-image-registry/image-registry-697d97f7c8-9vv2b: failed to check if pod openshift-image-registry/image-registry-697d97f7c8-9vv2b is in primary UDN: could not find OVN pod annotation in map[imageregistry.operator.openshift.io/dependencies-checksum:sha256:f0beb9378fd30968608b370e3877d4c76f7539f11a5eebf44bee42a8b2dd7068 openshift.io/required-scc:restricted-v2 openshift.io/scc:restricted-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nI1002 16:41:26.411476 6224 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf: failed to check if pod openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf is in primary UDN: could not find OVN pod annotation in map[openshift.io/required-scc:nonroot-v2 openshift.io/scc:nonroot-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nE1002 16:41:26.489404 6224 shared_informer.go:316] \\\\\\\"Unhandled Error\\\\\\\" err=\\\\\\\"unable to sync caches for ovn-lb-controller\\\\\\\" logger=\\\\\\\"UnhandledError\\\\\\\"\\\\nI1002 16:41:26.490615 6224 ovnkube.go:599] Stopped ovnkube\\\\nI1002 16:41:26.490664 6224 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:41Z\\\",\\\"message\\\":\\\" failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z]\\\\nI1002 16:41:41.372050 6653 services_controller.go:434] Service openshift-machine-config-operator/machine-config-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-config-operator openshift-machine-config-operator 8bc1afc2-8724-4135-84df-aee09f23af4c 4514 0 2025-02-23 05:12:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:machine-config-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:mco-proxy-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00067dfeb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Na\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.133711 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"560139b3-154f-45d4-ae7d-db62351531c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07b9557dcbd29f15e9ae34cc712160a7f5f34330deaa1c08774efd54e68c4ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28b9ee75280e7e9c44e99c1b82c9760990628488cb359247df6483530fa09ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc88fc50e88cd95a703f74122351a86eee2dc44b402e48893690ecd79c9e5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.149146 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.158867 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.158908 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.158951 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.158978 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.158994 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:43Z","lastTransitionTime":"2025-10-02T16:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.163534 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.179963 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.193449 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2x552" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e40838-d098-402e-b99a-819ce5a5977c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2x552\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.208343 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.223579 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.238739 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.255451 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.262349 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.262404 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.262419 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.262440 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.262456 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:43Z","lastTransitionTime":"2025-10-02T16:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.270480 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f7426ac250fc0c8a7afe7b809d6ba62dc69ae4ff0f2e595027f86958d505ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ef5f79f90831cb32131db7ea8d2f305c3664bd9d06dd98c007b74378f1bf1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.287987 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.318116 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.334750 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:42Z\\\",\\\"message\\\":\\\"2025-10-02T16:40:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_15c55b75-b1a1-40e0-876d-f4ed62e46169\\\\n2025-10-02T16:40:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_15c55b75-b1a1-40e0-876d-f4ed62e46169 to /host/opt/cni/bin/\\\\n2025-10-02T16:40:57Z [verbose] multus-daemon started\\\\n2025-10-02T16:40:57Z [verbose] Readiness Indicator file check\\\\n2025-10-02T16:41:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.353125 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:41Z\\\",\\\"message\\\":\\\" failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z]\\\\nI1002 16:41:41.372050 6653 services_controller.go:434] Service openshift-machine-config-operator/machine-config-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-config-operator openshift-machine-config-operator 8bc1afc2-8724-4135-84df-aee09f23af4c 4514 0 2025-02-23 05:12:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:machine-config-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:mco-proxy-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00067dfeb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Na\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zm46w_openshift-ovn-kubernetes(e441e501-09ee-4e3d-a763-c853de09a02c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.365165 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.365223 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.365259 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.365283 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.365295 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:43Z","lastTransitionTime":"2025-10-02T16:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.367371 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.382387 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.397611 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.412854 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.428026 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.442345 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.460065 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.468821 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.468919 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.468983 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.469011 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.469079 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:43Z","lastTransitionTime":"2025-10-02T16:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.472189 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2x552" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e40838-d098-402e-b99a-819ce5a5977c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2x552\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.484878 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"560139b3-154f-45d4-ae7d-db62351531c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07b9557dcbd29f15e9ae34cc712160a7f5f34330deaa1c08774efd54e68c4ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28b9ee75280e7e9c44e99c1b82c9760990628488cb359247df6483530fa09ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc88fc50e88cd95a703f74122351a86eee2dc44b402e48893690ecd79c9e5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.501133 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.517279 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.534352 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.545912 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f7426ac250fc0c8a7afe7b809d6ba62dc69ae4ff0f2e595027f86958d505ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ef5f79f90831cb32131db7ea8d2f305c3664bd9d06dd98c007b74378f1bf1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.559644 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.573462 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.573511 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.573524 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.573542 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.573556 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:43Z","lastTransitionTime":"2025-10-02T16:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.584224 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.598532 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:42Z\\\",\\\"message\\\":\\\"2025-10-02T16:40:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_15c55b75-b1a1-40e0-876d-f4ed62e46169\\\\n2025-10-02T16:40:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_15c55b75-b1a1-40e0-876d-f4ed62e46169 to /host/opt/cni/bin/\\\\n2025-10-02T16:40:57Z [verbose] multus-daemon started\\\\n2025-10-02T16:40:57Z [verbose] Readiness Indicator file check\\\\n2025-10-02T16:41:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.616254 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:41Z\\\",\\\"message\\\":\\\" failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z]\\\\nI1002 16:41:41.372050 6653 services_controller.go:434] Service openshift-machine-config-operator/machine-config-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-config-operator openshift-machine-config-operator 8bc1afc2-8724-4135-84df-aee09f23af4c 4514 0 2025-02-23 05:12:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:machine-config-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:mco-proxy-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00067dfeb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Na\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zm46w_openshift-ovn-kubernetes(e441e501-09ee-4e3d-a763-c853de09a02c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.632422 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.646911 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.659510 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.674021 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.676894 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.677116 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.677311 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.677492 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.677660 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:43Z","lastTransitionTime":"2025-10-02T16:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.689496 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.781361 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.781858 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.782041 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.782227 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.782439 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:43Z","lastTransitionTime":"2025-10-02T16:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.860102 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7d9g4_8871f396-b980-46a8-9477-b3be4889638f/kube-multus/0.log" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.860181 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7d9g4" event={"ID":"8871f396-b980-46a8-9477-b3be4889638f","Type":"ContainerStarted","Data":"19f7576c6dc4ec19e5895eb22a1790de5928dbd0c4779a000331f3f45df532c1"} Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.875663 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.885686 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.885742 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.885751 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.885768 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.885780 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:43Z","lastTransitionTime":"2025-10-02T16:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.892071 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.907649 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.921787 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.936995 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"560139b3-154f-45d4-ae7d-db62351531c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07b9557dcbd29f15e9ae34cc712160a7f5f34330deaa1c08774efd54e68c4ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28b9ee75280e7e9c44e99c1b82c9760990628488cb359247df6483530fa09ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc88fc50e88cd95a703f74122351a86eee2dc44b402e48893690ecd79c9e5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.952932 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.965714 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.979438 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.991506 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.991743 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.991834 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.991982 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.992098 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:43Z","lastTransitionTime":"2025-10-02T16:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:43 crc kubenswrapper[4808]: I1002 16:41:43.991858 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2x552" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e40838-d098-402e-b99a-819ce5a5977c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2x552\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.007386 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:44Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.022885 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:44Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.033047 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:44Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.050148 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:44Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.064498 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f7426ac250fc0c8a7afe7b809d6ba62dc69ae4ff0f2e595027f86958d505ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ef5f79f90831cb32131db7ea8d2f305c3664bd9d06dd98c007b74378f1bf1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:44Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.084001 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:44Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.095152 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.095434 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.095581 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.095761 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.095901 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:44Z","lastTransitionTime":"2025-10-02T16:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.107594 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:44Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.126090 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19f7576c6dc4ec19e5895eb22a1790de5928dbd0c4779a000331f3f45df532c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:42Z\\\",\\\"message\\\":\\\"2025-10-02T16:40:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_15c55b75-b1a1-40e0-876d-f4ed62e46169\\\\n2025-10-02T16:40:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_15c55b75-b1a1-40e0-876d-f4ed62e46169 to /host/opt/cni/bin/\\\\n2025-10-02T16:40:57Z [verbose] multus-daemon started\\\\n2025-10-02T16:40:57Z [verbose] Readiness Indicator file check\\\\n2025-10-02T16:41:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:44Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.157460 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:41Z\\\",\\\"message\\\":\\\" failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z]\\\\nI1002 16:41:41.372050 6653 services_controller.go:434] Service openshift-machine-config-operator/machine-config-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-config-operator openshift-machine-config-operator 8bc1afc2-8724-4135-84df-aee09f23af4c 4514 0 2025-02-23 05:12:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:machine-config-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:mco-proxy-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00067dfeb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Na\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zm46w_openshift-ovn-kubernetes(e441e501-09ee-4e3d-a763-c853de09a02c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:44Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.199330 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.199391 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.199409 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.199436 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.199456 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:44Z","lastTransitionTime":"2025-10-02T16:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.308911 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.309026 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.309056 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.309090 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.309118 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:44Z","lastTransitionTime":"2025-10-02T16:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.395625 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.395699 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.395767 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:44 crc kubenswrapper[4808]: E1002 16:41:44.395829 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.395641 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:44 crc kubenswrapper[4808]: E1002 16:41:44.395921 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:44 crc kubenswrapper[4808]: E1002 16:41:44.396009 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:44 crc kubenswrapper[4808]: E1002 16:41:44.396135 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.413247 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.413300 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.413311 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.413334 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.413346 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:44Z","lastTransitionTime":"2025-10-02T16:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.516000 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.516056 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.516067 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.516087 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.516100 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:44Z","lastTransitionTime":"2025-10-02T16:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.619125 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.619188 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.619200 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.619222 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.619259 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:44Z","lastTransitionTime":"2025-10-02T16:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.722960 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.723008 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.723021 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.723040 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.723052 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:44Z","lastTransitionTime":"2025-10-02T16:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.826393 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.826455 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.826466 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.826487 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.826503 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:44Z","lastTransitionTime":"2025-10-02T16:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.930192 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.930294 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.930313 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.930343 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:44 crc kubenswrapper[4808]: I1002 16:41:44.930362 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:44Z","lastTransitionTime":"2025-10-02T16:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.033489 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.033580 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.033598 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.033643 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.033663 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:45Z","lastTransitionTime":"2025-10-02T16:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.137861 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.137913 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.137925 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.137950 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.137964 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:45Z","lastTransitionTime":"2025-10-02T16:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.241539 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.241597 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.241610 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.241631 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.241644 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:45Z","lastTransitionTime":"2025-10-02T16:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.345601 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.345663 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.345676 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.345702 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.345714 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:45Z","lastTransitionTime":"2025-10-02T16:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.449089 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.449151 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.449174 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.449207 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.449229 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:45Z","lastTransitionTime":"2025-10-02T16:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.552999 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.553082 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.553105 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.553137 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.553162 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:45Z","lastTransitionTime":"2025-10-02T16:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.655940 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.656028 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.656058 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.656112 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.656141 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:45Z","lastTransitionTime":"2025-10-02T16:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.759653 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.759739 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.759748 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.759768 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.759779 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:45Z","lastTransitionTime":"2025-10-02T16:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.862918 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.862967 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.862978 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.862996 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.863008 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:45Z","lastTransitionTime":"2025-10-02T16:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.965267 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.965363 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.965382 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.965409 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:45 crc kubenswrapper[4808]: I1002 16:41:45.965426 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:45Z","lastTransitionTime":"2025-10-02T16:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.068492 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.068573 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.068594 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.068621 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.068642 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:46Z","lastTransitionTime":"2025-10-02T16:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.171575 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.171617 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.171629 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.171649 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.171662 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:46Z","lastTransitionTime":"2025-10-02T16:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.275131 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.275181 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.275196 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.275217 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.275250 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:46Z","lastTransitionTime":"2025-10-02T16:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.379756 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.379815 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.379835 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.379861 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.379880 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:46Z","lastTransitionTime":"2025-10-02T16:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.395257 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.395341 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.395398 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.395286 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:46 crc kubenswrapper[4808]: E1002 16:41:46.395479 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:46 crc kubenswrapper[4808]: E1002 16:41:46.395648 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:46 crc kubenswrapper[4808]: E1002 16:41:46.395695 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:46 crc kubenswrapper[4808]: E1002 16:41:46.395771 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.483142 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.483203 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.483213 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.483248 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.483259 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:46Z","lastTransitionTime":"2025-10-02T16:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.586478 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.586540 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.586551 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.586574 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.586587 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:46Z","lastTransitionTime":"2025-10-02T16:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.689846 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.689916 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.689930 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.689949 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.689962 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:46Z","lastTransitionTime":"2025-10-02T16:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.793416 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.793511 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.793576 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.793617 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.793643 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:46Z","lastTransitionTime":"2025-10-02T16:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.896482 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.896533 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.896543 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.896563 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.896574 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:46Z","lastTransitionTime":"2025-10-02T16:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.999378 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.999451 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.999468 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:46 crc kubenswrapper[4808]: I1002 16:41:46.999493 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:46.999511 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:46Z","lastTransitionTime":"2025-10-02T16:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.102046 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.102098 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.102107 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.102126 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.102140 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:47Z","lastTransitionTime":"2025-10-02T16:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.204975 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.205068 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.205090 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.205122 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.205140 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:47Z","lastTransitionTime":"2025-10-02T16:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.308255 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.308319 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.308338 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.308365 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.308385 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:47Z","lastTransitionTime":"2025-10-02T16:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.411317 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.411369 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.411378 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.411405 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.411417 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:47Z","lastTransitionTime":"2025-10-02T16:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.514591 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.514662 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.514681 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.514706 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.514728 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:47Z","lastTransitionTime":"2025-10-02T16:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.620718 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.620793 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.620809 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.620843 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.621448 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:47Z","lastTransitionTime":"2025-10-02T16:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.724769 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.724846 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.724857 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.724882 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.724897 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:47Z","lastTransitionTime":"2025-10-02T16:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.828029 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.828104 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.828118 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.828144 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.828160 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:47Z","lastTransitionTime":"2025-10-02T16:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.931450 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.931503 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.931516 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.931541 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:47 crc kubenswrapper[4808]: I1002 16:41:47.931554 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:47Z","lastTransitionTime":"2025-10-02T16:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.034654 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.034700 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.034710 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.034728 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.034740 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:48Z","lastTransitionTime":"2025-10-02T16:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.137176 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.137302 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.137319 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.137338 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.137355 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:48Z","lastTransitionTime":"2025-10-02T16:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.240329 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.240392 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.240404 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.240427 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.240439 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:48Z","lastTransitionTime":"2025-10-02T16:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.342917 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.342984 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.342996 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.343020 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.343036 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:48Z","lastTransitionTime":"2025-10-02T16:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.395667 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.395667 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.395652 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.395899 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:48 crc kubenswrapper[4808]: E1002 16:41:48.396066 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:48 crc kubenswrapper[4808]: E1002 16:41:48.396264 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:48 crc kubenswrapper[4808]: E1002 16:41:48.396554 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:48 crc kubenswrapper[4808]: E1002 16:41:48.396664 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.445876 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.445932 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.445947 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.445967 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.445979 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:48Z","lastTransitionTime":"2025-10-02T16:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.548668 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.548717 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.548729 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.548748 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.548760 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:48Z","lastTransitionTime":"2025-10-02T16:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.652516 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.652566 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.652577 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.652599 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.652613 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:48Z","lastTransitionTime":"2025-10-02T16:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.755838 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.755894 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.755904 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.755926 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.755940 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:48Z","lastTransitionTime":"2025-10-02T16:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.858797 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.858870 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.858886 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.858915 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.858932 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:48Z","lastTransitionTime":"2025-10-02T16:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.962252 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.962307 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.962319 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.962338 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:48 crc kubenswrapper[4808]: I1002 16:41:48.962351 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:48Z","lastTransitionTime":"2025-10-02T16:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.065593 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.065656 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.065666 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.065686 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.065704 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:49Z","lastTransitionTime":"2025-10-02T16:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.168577 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.168634 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.168646 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.168670 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.168683 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:49Z","lastTransitionTime":"2025-10-02T16:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.272385 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.272444 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.272454 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.272474 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.272488 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:49Z","lastTransitionTime":"2025-10-02T16:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.375786 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.375849 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.375860 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.375878 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.375891 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:49Z","lastTransitionTime":"2025-10-02T16:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.479807 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.479869 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.479880 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.479901 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.479914 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:49Z","lastTransitionTime":"2025-10-02T16:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.583607 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.583669 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.583683 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.583703 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.583715 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:49Z","lastTransitionTime":"2025-10-02T16:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.687374 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.687605 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.687618 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.687639 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.687653 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:49Z","lastTransitionTime":"2025-10-02T16:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.791121 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.791197 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.791211 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.791242 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.791255 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:49Z","lastTransitionTime":"2025-10-02T16:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.894175 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.894278 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.894298 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.894367 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.894402 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:49Z","lastTransitionTime":"2025-10-02T16:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.970675 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.970735 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.970749 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.970773 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.970790 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:49Z","lastTransitionTime":"2025-10-02T16:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:49 crc kubenswrapper[4808]: E1002 16:41:49.986220 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:49Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.992047 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.992097 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.992108 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.992130 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:49 crc kubenswrapper[4808]: I1002 16:41:49.992142 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:49Z","lastTransitionTime":"2025-10-02T16:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:50 crc kubenswrapper[4808]: E1002 16:41:50.006540 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:50Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.010971 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.011005 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.011018 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.011039 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.011050 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:50Z","lastTransitionTime":"2025-10-02T16:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:50 crc kubenswrapper[4808]: E1002 16:41:50.024632 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:50Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.029435 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.029486 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.029496 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.029520 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.029533 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:50Z","lastTransitionTime":"2025-10-02T16:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:50 crc kubenswrapper[4808]: E1002 16:41:50.045043 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:50Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.051019 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.051054 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.051065 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.051085 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.051098 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:50Z","lastTransitionTime":"2025-10-02T16:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:50 crc kubenswrapper[4808]: E1002 16:41:50.068017 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:50Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:50 crc kubenswrapper[4808]: E1002 16:41:50.068165 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.070822 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.070889 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.070922 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.070943 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.070954 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:50Z","lastTransitionTime":"2025-10-02T16:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.173939 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.173999 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.174010 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.174033 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.174043 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:50Z","lastTransitionTime":"2025-10-02T16:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.277625 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.277679 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.277693 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.277715 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.277725 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:50Z","lastTransitionTime":"2025-10-02T16:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.380211 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.380299 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.380312 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.380337 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.380355 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:50Z","lastTransitionTime":"2025-10-02T16:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.395728 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.395758 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.395837 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:50 crc kubenswrapper[4808]: E1002 16:41:50.395894 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.395952 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:50 crc kubenswrapper[4808]: E1002 16:41:50.396126 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:50 crc kubenswrapper[4808]: E1002 16:41:50.396291 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:50 crc kubenswrapper[4808]: E1002 16:41:50.396452 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.483745 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.483807 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.483819 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.483843 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.483856 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:50Z","lastTransitionTime":"2025-10-02T16:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.586687 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.586735 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.586745 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.586762 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.586773 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:50Z","lastTransitionTime":"2025-10-02T16:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.690225 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.690309 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.690319 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.690336 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.690346 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:50Z","lastTransitionTime":"2025-10-02T16:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.794209 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.794276 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.794294 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.794319 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.794335 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:50Z","lastTransitionTime":"2025-10-02T16:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.897263 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.897330 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.897341 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.897362 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:50 crc kubenswrapper[4808]: I1002 16:41:50.897377 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:50Z","lastTransitionTime":"2025-10-02T16:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.000613 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.000662 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.000677 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.000702 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.000718 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:51Z","lastTransitionTime":"2025-10-02T16:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.104927 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.104998 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.105018 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.105043 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.105093 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:51Z","lastTransitionTime":"2025-10-02T16:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.212597 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.212660 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.212697 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.212728 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.212753 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:51Z","lastTransitionTime":"2025-10-02T16:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.317198 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.317282 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.317295 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.317317 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.317331 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:51Z","lastTransitionTime":"2025-10-02T16:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.421692 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.421805 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.421821 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.421842 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.421855 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:51Z","lastTransitionTime":"2025-10-02T16:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.525007 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.525076 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.525085 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.525110 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.525126 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:51Z","lastTransitionTime":"2025-10-02T16:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.628614 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.628691 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.628701 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.628720 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.628732 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:51Z","lastTransitionTime":"2025-10-02T16:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.732980 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.733568 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.733576 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.733596 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.733606 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:51Z","lastTransitionTime":"2025-10-02T16:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.837245 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.837319 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.837330 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.837352 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.837366 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:51Z","lastTransitionTime":"2025-10-02T16:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.941632 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.941689 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.941700 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.941717 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:51 crc kubenswrapper[4808]: I1002 16:41:51.941729 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:51Z","lastTransitionTime":"2025-10-02T16:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.045631 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.045720 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.045744 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.045770 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.045788 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:52Z","lastTransitionTime":"2025-10-02T16:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.149832 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.149905 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.149916 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.149937 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.149982 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:52Z","lastTransitionTime":"2025-10-02T16:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.253143 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.253214 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.253224 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.253263 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.253277 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:52Z","lastTransitionTime":"2025-10-02T16:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.356202 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.356286 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.356297 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.356317 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.356332 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:52Z","lastTransitionTime":"2025-10-02T16:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.395183 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:52 crc kubenswrapper[4808]: E1002 16:41:52.395399 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.395472 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:52 crc kubenswrapper[4808]: E1002 16:41:52.395517 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.395563 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:52 crc kubenswrapper[4808]: E1002 16:41:52.395609 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.395645 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:52 crc kubenswrapper[4808]: E1002 16:41:52.395688 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.459319 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.459388 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.459409 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.459439 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.459457 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:52Z","lastTransitionTime":"2025-10-02T16:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.563388 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.563462 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.563480 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.563508 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.563530 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:52Z","lastTransitionTime":"2025-10-02T16:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.666987 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.667068 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.667086 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.667133 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.667151 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:52Z","lastTransitionTime":"2025-10-02T16:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.770585 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.770653 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.770674 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.770702 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.770722 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:52Z","lastTransitionTime":"2025-10-02T16:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.873842 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.873912 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.873923 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.873945 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.873957 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:52Z","lastTransitionTime":"2025-10-02T16:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.977221 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.977293 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.977305 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.977322 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:52 crc kubenswrapper[4808]: I1002 16:41:52.977334 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:52Z","lastTransitionTime":"2025-10-02T16:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.080681 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.080791 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.080811 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.080846 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.080867 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:53Z","lastTransitionTime":"2025-10-02T16:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.185405 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.185473 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.185483 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.185505 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.185517 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:53Z","lastTransitionTime":"2025-10-02T16:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.289608 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.289744 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.289771 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.289828 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.289852 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:53Z","lastTransitionTime":"2025-10-02T16:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.393416 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.393494 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.393511 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.393547 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.393567 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:53Z","lastTransitionTime":"2025-10-02T16:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.426197 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.442145 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2x552" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e40838-d098-402e-b99a-819ce5a5977c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2x552\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.461926 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"560139b3-154f-45d4-ae7d-db62351531c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07b9557dcbd29f15e9ae34cc712160a7f5f34330deaa1c08774efd54e68c4ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28b9ee75280e7e9c44e99c1b82c9760990628488cb359247df6483530fa09ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc88fc50e88cd95a703f74122351a86eee2dc44b402e48893690ecd79c9e5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.481041 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.498039 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.498103 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.498121 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.498145 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.498163 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:53Z","lastTransitionTime":"2025-10-02T16:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.505432 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.525285 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.545662 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f7426ac250fc0c8a7afe7b809d6ba62dc69ae4ff0f2e595027f86958d505ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ef5f79f90831cb32131db7ea8d2f305c3664bd9d06dd98c007b74378f1bf1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.563983 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.585984 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.601863 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.601942 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.602157 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.602212 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.602291 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:53Z","lastTransitionTime":"2025-10-02T16:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.605585 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.627060 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19f7576c6dc4ec19e5895eb22a1790de5928dbd0c4779a000331f3f45df532c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:42Z\\\",\\\"message\\\":\\\"2025-10-02T16:40:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_15c55b75-b1a1-40e0-876d-f4ed62e46169\\\\n2025-10-02T16:40:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_15c55b75-b1a1-40e0-876d-f4ed62e46169 to /host/opt/cni/bin/\\\\n2025-10-02T16:40:57Z [verbose] multus-daemon started\\\\n2025-10-02T16:40:57Z [verbose] Readiness Indicator file check\\\\n2025-10-02T16:41:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.650439 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:41Z\\\",\\\"message\\\":\\\" failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z]\\\\nI1002 16:41:41.372050 6653 services_controller.go:434] Service openshift-machine-config-operator/machine-config-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-config-operator openshift-machine-config-operator 8bc1afc2-8724-4135-84df-aee09f23af4c 4514 0 2025-02-23 05:12:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:machine-config-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:mco-proxy-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00067dfeb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Na\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zm46w_openshift-ovn-kubernetes(e441e501-09ee-4e3d-a763-c853de09a02c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.675353 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.696525 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.705944 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.705986 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.706003 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.706027 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.706037 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:53Z","lastTransitionTime":"2025-10-02T16:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.712638 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.731999 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.753889 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.770772 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.809305 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.809372 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.809382 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.809401 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.809416 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:53Z","lastTransitionTime":"2025-10-02T16:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.912920 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.912961 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.912969 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.912986 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:53 crc kubenswrapper[4808]: I1002 16:41:53.912996 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:53Z","lastTransitionTime":"2025-10-02T16:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.015878 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.015957 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.015974 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.016000 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.016020 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:54Z","lastTransitionTime":"2025-10-02T16:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.118816 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.118880 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.118897 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.118924 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.118942 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:54Z","lastTransitionTime":"2025-10-02T16:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.221830 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.221910 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.222019 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.222053 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.222076 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:54Z","lastTransitionTime":"2025-10-02T16:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.325304 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.325352 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.325369 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.325394 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.325411 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:54Z","lastTransitionTime":"2025-10-02T16:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.395688 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.395743 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.395759 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.395865 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:54 crc kubenswrapper[4808]: E1002 16:41:54.396071 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:54 crc kubenswrapper[4808]: E1002 16:41:54.396345 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:54 crc kubenswrapper[4808]: E1002 16:41:54.396441 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:54 crc kubenswrapper[4808]: E1002 16:41:54.396624 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.397907 4808 scope.go:117] "RemoveContainer" containerID="98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8" Oct 02 16:41:54 crc kubenswrapper[4808]: E1002 16:41:54.398225 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zm46w_openshift-ovn-kubernetes(e441e501-09ee-4e3d-a763-c853de09a02c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.429435 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.429499 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.429521 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.429549 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.429575 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:54Z","lastTransitionTime":"2025-10-02T16:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.533368 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.533440 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.533452 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.533469 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.533479 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:54Z","lastTransitionTime":"2025-10-02T16:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.636838 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.636894 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.636912 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.636937 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.637009 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:54Z","lastTransitionTime":"2025-10-02T16:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.740633 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.740696 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.740712 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.740738 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.740756 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:54Z","lastTransitionTime":"2025-10-02T16:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.843745 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.843800 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.843813 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.843831 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.843849 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:54Z","lastTransitionTime":"2025-10-02T16:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.946659 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.946711 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.946733 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.946758 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:54 crc kubenswrapper[4808]: I1002 16:41:54.946777 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:54Z","lastTransitionTime":"2025-10-02T16:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.050703 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.050761 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.050781 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.050805 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.050821 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:55Z","lastTransitionTime":"2025-10-02T16:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.153974 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.154104 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.154128 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.154153 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.154171 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:55Z","lastTransitionTime":"2025-10-02T16:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.258030 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.258097 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.258114 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.258180 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.258198 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:55Z","lastTransitionTime":"2025-10-02T16:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.360859 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.360937 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.360960 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.361048 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.361078 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:55Z","lastTransitionTime":"2025-10-02T16:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.412472 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.464464 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.464519 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.464535 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.464559 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.464577 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:55Z","lastTransitionTime":"2025-10-02T16:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.567765 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.567829 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.567849 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.567872 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.567889 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:55Z","lastTransitionTime":"2025-10-02T16:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.671398 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.671471 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.671494 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.671523 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.671545 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:55Z","lastTransitionTime":"2025-10-02T16:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.775348 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.775423 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.775440 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.775464 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.775482 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:55Z","lastTransitionTime":"2025-10-02T16:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.879350 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.879401 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.879422 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.879446 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.879465 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:55Z","lastTransitionTime":"2025-10-02T16:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.982280 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.982327 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.982340 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.982357 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:55 crc kubenswrapper[4808]: I1002 16:41:55.982374 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:55Z","lastTransitionTime":"2025-10-02T16:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.086101 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.086523 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.086739 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.086904 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.087025 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:56Z","lastTransitionTime":"2025-10-02T16:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.180293 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.180449 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:56 crc kubenswrapper[4808]: E1002 16:41:56.180601 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:00.180531361 +0000 UTC m=+147.506060401 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:41:56 crc kubenswrapper[4808]: E1002 16:41:56.180609 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.180714 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:56 crc kubenswrapper[4808]: E1002 16:41:56.180748 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 16:43:00.180726316 +0000 UTC m=+147.506255346 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 16:41:56 crc kubenswrapper[4808]: E1002 16:41:56.180902 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 16:41:56 crc kubenswrapper[4808]: E1002 16:41:56.180969 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 16:43:00.180952542 +0000 UTC m=+147.506481572 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.191604 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.191664 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.191677 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.191699 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.191713 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:56Z","lastTransitionTime":"2025-10-02T16:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.281944 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.282007 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:56 crc kubenswrapper[4808]: E1002 16:41:56.282142 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 16:41:56 crc kubenswrapper[4808]: E1002 16:41:56.282159 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 16:41:56 crc kubenswrapper[4808]: E1002 16:41:56.282170 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:41:56 crc kubenswrapper[4808]: E1002 16:41:56.282192 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 16:41:56 crc kubenswrapper[4808]: E1002 16:41:56.282230 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 16:41:56 crc kubenswrapper[4808]: E1002 16:41:56.282275 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:41:56 crc kubenswrapper[4808]: E1002 16:41:56.282227 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 16:43:00.282212107 +0000 UTC m=+147.607741107 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:41:56 crc kubenswrapper[4808]: E1002 16:41:56.282355 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 16:43:00.28233653 +0000 UTC m=+147.607865540 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.295066 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.295131 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.295153 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.295183 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.295208 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:56Z","lastTransitionTime":"2025-10-02T16:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.395416 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:56 crc kubenswrapper[4808]: E1002 16:41:56.395848 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.395473 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.396105 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:56 crc kubenswrapper[4808]: E1002 16:41:56.396136 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.395442 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:56 crc kubenswrapper[4808]: E1002 16:41:56.396334 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:56 crc kubenswrapper[4808]: E1002 16:41:56.396431 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.397873 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.397946 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.397972 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.398000 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.398023 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:56Z","lastTransitionTime":"2025-10-02T16:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.501774 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.501811 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.501820 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.501835 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.501845 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:56Z","lastTransitionTime":"2025-10-02T16:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.609795 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.609872 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.609891 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.609917 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.609944 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:56Z","lastTransitionTime":"2025-10-02T16:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.714534 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.714598 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.714611 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.714632 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.714647 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:56Z","lastTransitionTime":"2025-10-02T16:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.818365 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.818454 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.818478 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.818512 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.818538 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:56Z","lastTransitionTime":"2025-10-02T16:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.921734 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.921808 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.921827 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.921856 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:56 crc kubenswrapper[4808]: I1002 16:41:56.921874 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:56Z","lastTransitionTime":"2025-10-02T16:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.025799 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.025854 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.025864 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.025882 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.025893 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:57Z","lastTransitionTime":"2025-10-02T16:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.129323 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.129381 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.129398 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.129419 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.129431 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:57Z","lastTransitionTime":"2025-10-02T16:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.232494 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.232574 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.232588 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.232607 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.232620 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:57Z","lastTransitionTime":"2025-10-02T16:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.335795 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.335872 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.335889 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.335911 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.335921 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:57Z","lastTransitionTime":"2025-10-02T16:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.439493 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.439544 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.439561 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.439582 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.439596 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:57Z","lastTransitionTime":"2025-10-02T16:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.542099 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.542136 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.542144 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.542162 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.542172 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:57Z","lastTransitionTime":"2025-10-02T16:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.644505 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.644550 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.644561 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.644577 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.644618 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:57Z","lastTransitionTime":"2025-10-02T16:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.748198 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.748269 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.748286 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.748308 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.748322 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:57Z","lastTransitionTime":"2025-10-02T16:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.850891 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.850961 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.850974 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.850994 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.851006 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:57Z","lastTransitionTime":"2025-10-02T16:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.953395 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.953444 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.953455 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.953473 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:57 crc kubenswrapper[4808]: I1002 16:41:57.953488 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:57Z","lastTransitionTime":"2025-10-02T16:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.056412 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.056453 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.056464 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.056479 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.056491 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:58Z","lastTransitionTime":"2025-10-02T16:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.159638 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.160041 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.160543 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.160645 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.160714 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:58Z","lastTransitionTime":"2025-10-02T16:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.263909 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.263966 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.263977 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.263997 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.264010 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:58Z","lastTransitionTime":"2025-10-02T16:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.367625 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.367712 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.367731 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.367769 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.367791 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:58Z","lastTransitionTime":"2025-10-02T16:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.395533 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.395552 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.395560 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.395732 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:41:58 crc kubenswrapper[4808]: E1002 16:41:58.395900 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:41:58 crc kubenswrapper[4808]: E1002 16:41:58.395995 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:41:58 crc kubenswrapper[4808]: E1002 16:41:58.396132 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:41:58 crc kubenswrapper[4808]: E1002 16:41:58.396304 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.470993 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.471045 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.471059 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.471078 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.471092 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:58Z","lastTransitionTime":"2025-10-02T16:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.574244 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.575385 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.575631 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.575736 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.575811 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:58Z","lastTransitionTime":"2025-10-02T16:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.679557 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.679630 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.679651 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.679681 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.679829 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:58Z","lastTransitionTime":"2025-10-02T16:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.782894 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.782932 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.782942 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.782964 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.782978 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:58Z","lastTransitionTime":"2025-10-02T16:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.885140 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.885175 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.885186 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.885203 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.885215 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:58Z","lastTransitionTime":"2025-10-02T16:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.988422 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.988467 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.988478 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.988496 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:58 crc kubenswrapper[4808]: I1002 16:41:58.988507 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:58Z","lastTransitionTime":"2025-10-02T16:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.090683 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.090723 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.090734 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.090749 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.090761 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:59Z","lastTransitionTime":"2025-10-02T16:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.193969 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.194031 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.194060 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.194086 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.194104 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:59Z","lastTransitionTime":"2025-10-02T16:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.297524 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.297657 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.297683 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.297713 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.297733 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:59Z","lastTransitionTime":"2025-10-02T16:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.400473 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.400532 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.400546 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.400566 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.400578 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:59Z","lastTransitionTime":"2025-10-02T16:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.503562 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.503645 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.503659 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.503680 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.503692 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:59Z","lastTransitionTime":"2025-10-02T16:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.607469 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.607529 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.607542 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.607561 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.607573 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:59Z","lastTransitionTime":"2025-10-02T16:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.710390 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.710453 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.710463 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.710483 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.710496 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:59Z","lastTransitionTime":"2025-10-02T16:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.813426 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.813466 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.813476 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.813490 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.813501 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:59Z","lastTransitionTime":"2025-10-02T16:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.916265 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.916588 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.916683 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.917197 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:41:59 crc kubenswrapper[4808]: I1002 16:41:59.917301 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:41:59Z","lastTransitionTime":"2025-10-02T16:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.020002 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.020045 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.020056 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.020074 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.020085 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:00Z","lastTransitionTime":"2025-10-02T16:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.122962 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.123023 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.123036 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.123063 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.123079 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:00Z","lastTransitionTime":"2025-10-02T16:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.160131 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.160179 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.160190 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.160210 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.160222 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:00Z","lastTransitionTime":"2025-10-02T16:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:00 crc kubenswrapper[4808]: E1002 16:42:00.175793 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:00Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.180038 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.180097 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.180115 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.180145 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.180166 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:00Z","lastTransitionTime":"2025-10-02T16:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:00 crc kubenswrapper[4808]: E1002 16:42:00.200038 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:00Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.205338 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.205390 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.205401 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.205419 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.205430 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:00Z","lastTransitionTime":"2025-10-02T16:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:00 crc kubenswrapper[4808]: E1002 16:42:00.222414 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:00Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.227019 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.227051 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.227061 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.227077 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.227087 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:00Z","lastTransitionTime":"2025-10-02T16:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:00 crc kubenswrapper[4808]: E1002 16:42:00.243431 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:00Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.248372 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.248476 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.248496 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.248571 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.248595 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:00Z","lastTransitionTime":"2025-10-02T16:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:00 crc kubenswrapper[4808]: E1002 16:42:00.264656 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:00Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:00 crc kubenswrapper[4808]: E1002 16:42:00.264801 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.268002 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.268077 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.268107 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.268148 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.268174 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:00Z","lastTransitionTime":"2025-10-02T16:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.370720 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.370764 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.370777 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.370794 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.370808 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:00Z","lastTransitionTime":"2025-10-02T16:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.395131 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:00 crc kubenswrapper[4808]: E1002 16:42:00.395222 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.395134 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.395279 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.395337 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:00 crc kubenswrapper[4808]: E1002 16:42:00.395362 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:00 crc kubenswrapper[4808]: E1002 16:42:00.395392 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:00 crc kubenswrapper[4808]: E1002 16:42:00.395437 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.473445 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.473494 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.473511 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.473531 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.473547 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:00Z","lastTransitionTime":"2025-10-02T16:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.576631 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.576669 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.576678 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.576691 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.576701 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:00Z","lastTransitionTime":"2025-10-02T16:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.680078 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.680154 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.680185 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.680357 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.680389 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:00Z","lastTransitionTime":"2025-10-02T16:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.783282 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.783329 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.783341 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.783358 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.783370 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:00Z","lastTransitionTime":"2025-10-02T16:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.886374 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.886428 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.886445 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.886471 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.886489 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:00Z","lastTransitionTime":"2025-10-02T16:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.989570 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.989639 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.989658 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.989683 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:00 crc kubenswrapper[4808]: I1002 16:42:00.989703 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:00Z","lastTransitionTime":"2025-10-02T16:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.093225 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.093351 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.093365 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.093386 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.093401 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:01Z","lastTransitionTime":"2025-10-02T16:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.196727 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.196803 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.196822 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.196849 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.196866 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:01Z","lastTransitionTime":"2025-10-02T16:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.300148 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.300220 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.300253 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.300273 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.300284 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:01Z","lastTransitionTime":"2025-10-02T16:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.402660 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.402732 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.402757 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.402787 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.402811 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:01Z","lastTransitionTime":"2025-10-02T16:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.505490 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.505758 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.505770 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.505791 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.505808 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:01Z","lastTransitionTime":"2025-10-02T16:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.608579 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.608645 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.608664 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.608686 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.608701 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:01Z","lastTransitionTime":"2025-10-02T16:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.711683 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.711746 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.711769 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.711797 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.711819 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:01Z","lastTransitionTime":"2025-10-02T16:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.814658 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.814736 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.814755 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.814780 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.814802 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:01Z","lastTransitionTime":"2025-10-02T16:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.918191 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.918307 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.918328 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.918355 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:01 crc kubenswrapper[4808]: I1002 16:42:01.918374 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:01Z","lastTransitionTime":"2025-10-02T16:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.021805 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.021860 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.021874 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.021895 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.021911 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:02Z","lastTransitionTime":"2025-10-02T16:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.124664 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.124736 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.124757 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.124781 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.124800 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:02Z","lastTransitionTime":"2025-10-02T16:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.228598 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.228647 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.228657 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.228676 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.228687 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:02Z","lastTransitionTime":"2025-10-02T16:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.331873 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.331927 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.331945 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.331970 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.331990 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:02Z","lastTransitionTime":"2025-10-02T16:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.396079 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.396190 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.396210 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.396304 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:02 crc kubenswrapper[4808]: E1002 16:42:02.396384 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:02 crc kubenswrapper[4808]: E1002 16:42:02.396552 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:02 crc kubenswrapper[4808]: E1002 16:42:02.396745 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:02 crc kubenswrapper[4808]: E1002 16:42:02.396892 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.435750 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.435803 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.435812 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.435830 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.435843 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:02Z","lastTransitionTime":"2025-10-02T16:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.539022 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.539087 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.539103 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.539127 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.539141 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:02Z","lastTransitionTime":"2025-10-02T16:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.642505 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.642548 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.642563 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.642583 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.642598 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:02Z","lastTransitionTime":"2025-10-02T16:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.745442 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.745510 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.745528 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.745554 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.745572 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:02Z","lastTransitionTime":"2025-10-02T16:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.848665 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.848712 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.848732 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.848758 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.848775 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:02Z","lastTransitionTime":"2025-10-02T16:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.951823 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.951876 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.951886 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.951904 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:02 crc kubenswrapper[4808]: I1002 16:42:02.951917 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:02Z","lastTransitionTime":"2025-10-02T16:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.054599 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.054659 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.054675 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.054698 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.054714 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:03Z","lastTransitionTime":"2025-10-02T16:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.158217 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.158293 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.158310 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.158331 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.158343 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:03Z","lastTransitionTime":"2025-10-02T16:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.262105 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.262212 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.262224 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.262663 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.262675 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:03Z","lastTransitionTime":"2025-10-02T16:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.365435 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.365495 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.365508 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.365530 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.365543 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:03Z","lastTransitionTime":"2025-10-02T16:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.412614 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19f7576c6dc4ec19e5895eb22a1790de5928dbd0c4779a000331f3f45df532c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:42Z\\\",\\\"message\\\":\\\"2025-10-02T16:40:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_15c55b75-b1a1-40e0-876d-f4ed62e46169\\\\n2025-10-02T16:40:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_15c55b75-b1a1-40e0-876d-f4ed62e46169 to /host/opt/cni/bin/\\\\n2025-10-02T16:40:57Z [verbose] multus-daemon started\\\\n2025-10-02T16:40:57Z [verbose] Readiness Indicator file check\\\\n2025-10-02T16:41:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.442322 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:41Z\\\",\\\"message\\\":\\\" failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z]\\\\nI1002 16:41:41.372050 6653 services_controller.go:434] Service openshift-machine-config-operator/machine-config-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-config-operator openshift-machine-config-operator 8bc1afc2-8724-4135-84df-aee09f23af4c 4514 0 2025-02-23 05:12:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:machine-config-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:mco-proxy-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00067dfeb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Na\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zm46w_openshift-ovn-kubernetes(e441e501-09ee-4e3d-a763-c853de09a02c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.458901 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.468331 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.468367 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.468378 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.468396 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.468410 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:03Z","lastTransitionTime":"2025-10-02T16:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.472783 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934256ab-6db0-4440-971e-fea9bd9ce459\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe995b6f7e46579fae83eb2918d20e9801ed0f52df4753ad1871c41b9e11abc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a254a3cf9d11b54b681fc4cda92b06f88ff70b87906c439cdaecc58c7b0738f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a254a3cf9d11b54b681fc4cda92b06f88ff70b87906c439cdaecc58c7b0738f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.492885 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.512459 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.529953 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.547112 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.561931 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.571499 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.571563 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.571572 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.571592 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.571604 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:03Z","lastTransitionTime":"2025-10-02T16:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.575371 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.587022 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2x552" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e40838-d098-402e-b99a-819ce5a5977c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2x552\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.597945 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"560139b3-154f-45d4-ae7d-db62351531c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07b9557dcbd29f15e9ae34cc712160a7f5f34330deaa1c08774efd54e68c4ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28b9ee75280e7e9c44e99c1b82c9760990628488cb359247df6483530fa09ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc88fc50e88cd95a703f74122351a86eee2dc44b402e48893690ecd79c9e5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.611093 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.634783 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.661979 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.673992 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.674032 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.674043 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.674060 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.674071 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:03Z","lastTransitionTime":"2025-10-02T16:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.681041 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f7426ac250fc0c8a7afe7b809d6ba62dc69ae4ff0f2e595027f86958d505ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ef5f79f90831cb32131db7ea8d2f305c3664bd9d06dd98c007b74378f1bf1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.697988 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.714407 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.726515 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.777187 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.777361 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.777443 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.777531 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.777611 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:03Z","lastTransitionTime":"2025-10-02T16:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.880801 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.881167 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.881288 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.881383 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.881462 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:03Z","lastTransitionTime":"2025-10-02T16:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.984204 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.984270 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.984283 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.984301 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:03 crc kubenswrapper[4808]: I1002 16:42:03.984312 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:03Z","lastTransitionTime":"2025-10-02T16:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.087103 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.087165 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.087179 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.087203 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.087217 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:04Z","lastTransitionTime":"2025-10-02T16:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.190426 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.190802 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.190895 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.190998 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.191071 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:04Z","lastTransitionTime":"2025-10-02T16:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.293899 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.293937 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.293949 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.293966 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.293977 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:04Z","lastTransitionTime":"2025-10-02T16:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.395360 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.395957 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.396084 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:04 crc kubenswrapper[4808]: E1002 16:42:04.396204 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:04 crc kubenswrapper[4808]: E1002 16:42:04.396138 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.396125 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:04 crc kubenswrapper[4808]: E1002 16:42:04.396376 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:04 crc kubenswrapper[4808]: E1002 16:42:04.396532 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.397531 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.397572 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.397585 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.397602 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.397616 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:04Z","lastTransitionTime":"2025-10-02T16:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.500957 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.501411 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.501490 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.501611 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.501691 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:04Z","lastTransitionTime":"2025-10-02T16:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.604139 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.604187 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.604198 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.604218 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.604260 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:04Z","lastTransitionTime":"2025-10-02T16:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.707742 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.707789 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.707802 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.707820 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.707832 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:04Z","lastTransitionTime":"2025-10-02T16:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.810309 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.810343 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.810352 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.810368 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.810378 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:04Z","lastTransitionTime":"2025-10-02T16:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.913117 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.913156 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.913168 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.913188 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:04 crc kubenswrapper[4808]: I1002 16:42:04.913199 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:04Z","lastTransitionTime":"2025-10-02T16:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.015509 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.015839 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.015908 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.015987 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.016056 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:05Z","lastTransitionTime":"2025-10-02T16:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.119818 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.119902 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.119922 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.119951 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.119972 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:05Z","lastTransitionTime":"2025-10-02T16:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.223204 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.223289 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.223299 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.223326 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.223337 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:05Z","lastTransitionTime":"2025-10-02T16:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.326761 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.326837 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.326850 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.326874 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.326888 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:05Z","lastTransitionTime":"2025-10-02T16:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.429440 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.429486 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.429496 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.429518 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.429528 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:05Z","lastTransitionTime":"2025-10-02T16:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.538253 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.538316 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.538328 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.538356 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.538373 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:05Z","lastTransitionTime":"2025-10-02T16:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.641337 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.641384 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.641398 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.641421 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.641437 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:05Z","lastTransitionTime":"2025-10-02T16:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.744642 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.744751 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.744788 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.744826 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.744850 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:05Z","lastTransitionTime":"2025-10-02T16:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.848456 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.848537 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.848547 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.848566 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.848577 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:05Z","lastTransitionTime":"2025-10-02T16:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.951960 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.952063 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.952093 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.952153 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:05 crc kubenswrapper[4808]: I1002 16:42:05.952176 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:05Z","lastTransitionTime":"2025-10-02T16:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.056833 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.056921 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.056945 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.056984 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.057007 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:06Z","lastTransitionTime":"2025-10-02T16:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.159927 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.160030 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.160045 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.160086 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.160101 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:06Z","lastTransitionTime":"2025-10-02T16:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.263739 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.263808 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.263821 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.263851 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.263875 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:06Z","lastTransitionTime":"2025-10-02T16:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.367325 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.367380 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.367400 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.367424 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.367440 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:06Z","lastTransitionTime":"2025-10-02T16:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.395817 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.395886 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.395945 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.396034 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:06 crc kubenswrapper[4808]: E1002 16:42:06.396025 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:06 crc kubenswrapper[4808]: E1002 16:42:06.396211 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:06 crc kubenswrapper[4808]: E1002 16:42:06.396311 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:06 crc kubenswrapper[4808]: E1002 16:42:06.396399 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.397085 4808 scope.go:117] "RemoveContainer" containerID="98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.471202 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.471283 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.471298 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.471319 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.471345 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:06Z","lastTransitionTime":"2025-10-02T16:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.575608 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.575667 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.575679 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.575699 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.575714 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:06Z","lastTransitionTime":"2025-10-02T16:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.678544 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.678595 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.678605 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.678624 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.678635 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:06Z","lastTransitionTime":"2025-10-02T16:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.781954 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.782006 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.782027 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.782050 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.782064 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:06Z","lastTransitionTime":"2025-10-02T16:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.885313 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.885369 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.885383 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.885403 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.885419 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:06Z","lastTransitionTime":"2025-10-02T16:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.945532 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zm46w_e441e501-09ee-4e3d-a763-c853de09a02c/ovnkube-controller/2.log" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.948487 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerStarted","Data":"f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97"} Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.949349 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.965630 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"560139b3-154f-45d4-ae7d-db62351531c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07b9557dcbd29f15e9ae34cc712160a7f5f34330deaa1c08774efd54e68c4ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28b9ee75280e7e9c44e99c1b82c9760990628488cb359247df6483530fa09ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc88fc50e88cd95a703f74122351a86eee2dc44b402e48893690ecd79c9e5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.980164 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.988828 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.988880 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.988890 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.988909 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.988921 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:06Z","lastTransitionTime":"2025-10-02T16:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:06 crc kubenswrapper[4808]: I1002 16:42:06.999206 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:06Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.016778 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.034117 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2x552" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e40838-d098-402e-b99a-819ce5a5977c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2x552\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.051759 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.067092 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.088913 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.091545 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.091592 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.091602 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.091621 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.091632 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:07Z","lastTransitionTime":"2025-10-02T16:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.123379 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.138335 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f7426ac250fc0c8a7afe7b809d6ba62dc69ae4ff0f2e595027f86958d505ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ef5f79f90831cb32131db7ea8d2f305c3664bd9d06dd98c007b74378f1bf1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.155622 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.168392 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934256ab-6db0-4440-971e-fea9bd9ce459\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe995b6f7e46579fae83eb2918d20e9801ed0f52df4753ad1871c41b9e11abc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a254a3cf9d11b54b681fc4cda92b06f88ff70b87906c439cdaecc58c7b0738f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a254a3cf9d11b54b681fc4cda92b06f88ff70b87906c439cdaecc58c7b0738f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.193534 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.194410 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.194446 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.194457 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.194474 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.194485 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:07Z","lastTransitionTime":"2025-10-02T16:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.210540 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19f7576c6dc4ec19e5895eb22a1790de5928dbd0c4779a000331f3f45df532c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:42Z\\\",\\\"message\\\":\\\"2025-10-02T16:40:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_15c55b75-b1a1-40e0-876d-f4ed62e46169\\\\n2025-10-02T16:40:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_15c55b75-b1a1-40e0-876d-f4ed62e46169 to /host/opt/cni/bin/\\\\n2025-10-02T16:40:57Z [verbose] multus-daemon started\\\\n2025-10-02T16:40:57Z [verbose] Readiness Indicator file check\\\\n2025-10-02T16:41:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.233215 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:41Z\\\",\\\"message\\\":\\\" failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z]\\\\nI1002 16:41:41.372050 6653 services_controller.go:434] Service openshift-machine-config-operator/machine-config-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-config-operator openshift-machine-config-operator 8bc1afc2-8724-4135-84df-aee09f23af4c 4514 0 2025-02-23 05:12:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:machine-config-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:mco-proxy-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00067dfeb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Na\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:42:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.248917 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.267910 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.281678 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.295305 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.297379 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.297620 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.297641 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.297663 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.297675 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:07Z","lastTransitionTime":"2025-10-02T16:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.401314 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.401384 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.401394 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.401422 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.401436 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:07Z","lastTransitionTime":"2025-10-02T16:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.503750 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.503820 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.503836 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.503861 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.503874 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:07Z","lastTransitionTime":"2025-10-02T16:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.606614 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.606656 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.606668 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.606685 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.606697 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:07Z","lastTransitionTime":"2025-10-02T16:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.710293 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.710339 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.710351 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.710368 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.710380 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:07Z","lastTransitionTime":"2025-10-02T16:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.813142 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.813182 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.813194 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.813214 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.813225 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:07Z","lastTransitionTime":"2025-10-02T16:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.916974 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.917030 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.917042 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.917064 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.917077 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:07Z","lastTransitionTime":"2025-10-02T16:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.953322 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zm46w_e441e501-09ee-4e3d-a763-c853de09a02c/ovnkube-controller/3.log" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.954195 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zm46w_e441e501-09ee-4e3d-a763-c853de09a02c/ovnkube-controller/2.log" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.963303 4808 generic.go:334] "Generic (PLEG): container finished" podID="e441e501-09ee-4e3d-a763-c853de09a02c" containerID="f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97" exitCode=1 Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.963406 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerDied","Data":"f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97"} Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.963499 4808 scope.go:117] "RemoveContainer" containerID="98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.964714 4808 scope.go:117] "RemoveContainer" containerID="f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97" Oct 02 16:42:07 crc kubenswrapper[4808]: E1002 16:42:07.964946 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zm46w_openshift-ovn-kubernetes(e441e501-09ee-4e3d-a763-c853de09a02c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.982589 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:07 crc kubenswrapper[4808]: I1002 16:42:07.996292 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:07Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.010171 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.020360 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.020408 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.020421 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.020448 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.020470 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:08Z","lastTransitionTime":"2025-10-02T16:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.023709 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.037370 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2x552" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e40838-d098-402e-b99a-819ce5a5977c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2x552\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.052037 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"560139b3-154f-45d4-ae7d-db62351531c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07b9557dcbd29f15e9ae34cc712160a7f5f34330deaa1c08774efd54e68c4ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28b9ee75280e7e9c44e99c1b82c9760990628488cb359247df6483530fa09ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc88fc50e88cd95a703f74122351a86eee2dc44b402e48893690ecd79c9e5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.066024 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.079347 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.093096 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.104925 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f7426ac250fc0c8a7afe7b809d6ba62dc69ae4ff0f2e595027f86958d505ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ef5f79f90831cb32131db7ea8d2f305c3664bd9d06dd98c007b74378f1bf1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.118102 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.123062 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.123100 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.123110 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.123127 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.123138 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:08Z","lastTransitionTime":"2025-10-02T16:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.134417 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.147458 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.162766 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.195370 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d3a9de6ff85dac7733e948c5cf248e0b293928609b2af3e8e749329c7608a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:41Z\\\",\\\"message\\\":\\\" failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:41:41Z is after 2025-08-24T17:21:41Z]\\\\nI1002 16:41:41.372050 6653 services_controller.go:434] Service openshift-machine-config-operator/machine-config-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-config-operator openshift-machine-config-operator 8bc1afc2-8724-4135-84df-aee09f23af4c 4514 0 2025-02-23 05:12:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:machine-config-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:mco-proxy-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00067dfeb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Na\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:42:07Z\\\",\\\"message\\\":\\\"6:42:07.304780 6956 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 16:42:07.305177 6956 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 16:42:07.305193 6956 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 16:42:07.305940 6956 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 16:42:07.305996 6956 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 16:42:07.306026 6956 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 16:42:07.306043 6956 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 16:42:07.306073 6956 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 16:42:07.306102 6956 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 16:42:07.306141 6956 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 16:42:07.306148 6956 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 16:42:07.306165 6956 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 16:42:07.306192 6956 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 16:42:07.306200 6956 factory.go:656] Stopping watch factory\\\\nI1002 16:42:07.306286 6956 ovnkube.go:599] Stopped ovnkube\\\\nI1002 16:42:07.306207 6956 handler.go:208] Removed *v1.Node event handler 7\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:42:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.215397 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.226888 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.226939 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.226956 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.226976 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.226990 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:08Z","lastTransitionTime":"2025-10-02T16:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.227790 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934256ab-6db0-4440-971e-fea9bd9ce459\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe995b6f7e46579fae83eb2918d20e9801ed0f52df4753ad1871c41b9e11abc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a254a3cf9d11b54b681fc4cda92b06f88ff70b87906c439cdaecc58c7b0738f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a254a3cf9d11b54b681fc4cda92b06f88ff70b87906c439cdaecc58c7b0738f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.247642 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.261616 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19f7576c6dc4ec19e5895eb22a1790de5928dbd0c4779a000331f3f45df532c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:42Z\\\",\\\"message\\\":\\\"2025-10-02T16:40:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_15c55b75-b1a1-40e0-876d-f4ed62e46169\\\\n2025-10-02T16:40:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_15c55b75-b1a1-40e0-876d-f4ed62e46169 to /host/opt/cni/bin/\\\\n2025-10-02T16:40:57Z [verbose] multus-daemon started\\\\n2025-10-02T16:40:57Z [verbose] Readiness Indicator file check\\\\n2025-10-02T16:41:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.330516 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.330585 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.330603 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.330630 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.330647 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:08Z","lastTransitionTime":"2025-10-02T16:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.396192 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:08 crc kubenswrapper[4808]: E1002 16:42:08.396415 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.396672 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:08 crc kubenswrapper[4808]: E1002 16:42:08.396728 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.396849 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:08 crc kubenswrapper[4808]: E1002 16:42:08.396908 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.397018 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:08 crc kubenswrapper[4808]: E1002 16:42:08.397068 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.433209 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.433284 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.433299 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.433320 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.433332 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:08Z","lastTransitionTime":"2025-10-02T16:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.536114 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.536154 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.536165 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.536184 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.536198 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:08Z","lastTransitionTime":"2025-10-02T16:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.639589 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.639641 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.639650 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.639672 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.639688 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:08Z","lastTransitionTime":"2025-10-02T16:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.742538 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.742588 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.742601 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.742626 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.742640 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:08Z","lastTransitionTime":"2025-10-02T16:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.845994 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.846049 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.846059 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.846079 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.846093 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:08Z","lastTransitionTime":"2025-10-02T16:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.948998 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.949046 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.949057 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.949075 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.949092 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:08Z","lastTransitionTime":"2025-10-02T16:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.969351 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zm46w_e441e501-09ee-4e3d-a763-c853de09a02c/ovnkube-controller/3.log" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.974349 4808 scope.go:117] "RemoveContainer" containerID="f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97" Oct 02 16:42:08 crc kubenswrapper[4808]: E1002 16:42:08.974719 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zm46w_openshift-ovn-kubernetes(e441e501-09ee-4e3d-a763-c853de09a02c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" Oct 02 16:42:08 crc kubenswrapper[4808]: I1002 16:42:08.996066 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:08Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.008021 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934256ab-6db0-4440-971e-fea9bd9ce459\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe995b6f7e46579fae83eb2918d20e9801ed0f52df4753ad1871c41b9e11abc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a254a3cf9d11b54b681fc4cda92b06f88ff70b87906c439cdaecc58c7b0738f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a254a3cf9d11b54b681fc4cda92b06f88ff70b87906c439cdaecc58c7b0738f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.031252 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.053307 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.053343 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.053353 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.053374 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.053392 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:09Z","lastTransitionTime":"2025-10-02T16:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.057885 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19f7576c6dc4ec19e5895eb22a1790de5928dbd0c4779a000331f3f45df532c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:42Z\\\",\\\"message\\\":\\\"2025-10-02T16:40:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_15c55b75-b1a1-40e0-876d-f4ed62e46169\\\\n2025-10-02T16:40:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_15c55b75-b1a1-40e0-876d-f4ed62e46169 to /host/opt/cni/bin/\\\\n2025-10-02T16:40:57Z [verbose] multus-daemon started\\\\n2025-10-02T16:40:57Z [verbose] Readiness Indicator file check\\\\n2025-10-02T16:41:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.089945 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:42:07Z\\\",\\\"message\\\":\\\"6:42:07.304780 6956 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 16:42:07.305177 6956 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 16:42:07.305193 6956 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 16:42:07.305940 6956 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 16:42:07.305996 6956 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 16:42:07.306026 6956 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 16:42:07.306043 6956 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 16:42:07.306073 6956 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 16:42:07.306102 6956 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 16:42:07.306141 6956 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 16:42:07.306148 6956 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 16:42:07.306165 6956 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 16:42:07.306192 6956 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 16:42:07.306200 6956 factory.go:656] Stopping watch factory\\\\nI1002 16:42:07.306286 6956 ovnkube.go:599] Stopped ovnkube\\\\nI1002 16:42:07.306207 6956 handler.go:208] Removed *v1.Node event handler 7\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:42:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zm46w_openshift-ovn-kubernetes(e441e501-09ee-4e3d-a763-c853de09a02c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.113751 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.132557 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.150001 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.156773 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.156853 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.156878 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.156911 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.156936 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:09Z","lastTransitionTime":"2025-10-02T16:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.169469 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.190707 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"560139b3-154f-45d4-ae7d-db62351531c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07b9557dcbd29f15e9ae34cc712160a7f5f34330deaa1c08774efd54e68c4ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28b9ee75280e7e9c44e99c1b82c9760990628488cb359247df6483530fa09ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc88fc50e88cd95a703f74122351a86eee2dc44b402e48893690ecd79c9e5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.217497 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.237999 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.255155 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.260285 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.260386 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.260411 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.260442 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.260462 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:09Z","lastTransitionTime":"2025-10-02T16:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.269626 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2x552" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e40838-d098-402e-b99a-819ce5a5977c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2x552\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.284347 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.302271 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.315743 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.339857 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.358750 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f7426ac250fc0c8a7afe7b809d6ba62dc69ae4ff0f2e595027f86958d505ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ef5f79f90831cb32131db7ea8d2f305c3664bd9d06dd98c007b74378f1bf1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:09Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.363132 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.363212 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.363267 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.363303 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.363331 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:09Z","lastTransitionTime":"2025-10-02T16:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.466665 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.466717 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.466728 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.466746 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.466756 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:09Z","lastTransitionTime":"2025-10-02T16:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.570372 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.570462 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.570490 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.570524 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.570552 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:09Z","lastTransitionTime":"2025-10-02T16:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.673717 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.673805 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.673849 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.673873 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.673887 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:09Z","lastTransitionTime":"2025-10-02T16:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.777736 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.777812 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.777825 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.777890 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.777910 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:09Z","lastTransitionTime":"2025-10-02T16:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.881336 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.881407 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.881424 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.881451 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.881463 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:09Z","lastTransitionTime":"2025-10-02T16:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.983979 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.984045 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.984065 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.984094 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:09 crc kubenswrapper[4808]: I1002 16:42:09.984111 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:09Z","lastTransitionTime":"2025-10-02T16:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.087630 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.087692 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.087709 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.087733 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.087750 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:10Z","lastTransitionTime":"2025-10-02T16:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.191773 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.191826 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.191836 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.191858 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.191870 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:10Z","lastTransitionTime":"2025-10-02T16:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.294636 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.294720 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.294740 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.294769 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.294785 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:10Z","lastTransitionTime":"2025-10-02T16:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.396067 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.396193 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.396418 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.396537 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:10 crc kubenswrapper[4808]: E1002 16:42:10.396559 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:10 crc kubenswrapper[4808]: E1002 16:42:10.396663 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:10 crc kubenswrapper[4808]: E1002 16:42:10.396447 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:10 crc kubenswrapper[4808]: E1002 16:42:10.397400 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.398352 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.398390 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.398402 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.398423 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.398436 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:10Z","lastTransitionTime":"2025-10-02T16:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.501815 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.501895 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.501920 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.501954 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.501978 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:10Z","lastTransitionTime":"2025-10-02T16:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.605666 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.605736 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.605754 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.605782 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.605806 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:10Z","lastTransitionTime":"2025-10-02T16:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.627933 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.627984 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.627994 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.628012 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.628028 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:10Z","lastTransitionTime":"2025-10-02T16:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:10 crc kubenswrapper[4808]: E1002 16:42:10.644720 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:10Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.650129 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.650180 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.650195 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.650219 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.650257 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:10Z","lastTransitionTime":"2025-10-02T16:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:10 crc kubenswrapper[4808]: E1002 16:42:10.665549 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:10Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.670346 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.670421 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.670455 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.670479 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.670495 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:10Z","lastTransitionTime":"2025-10-02T16:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:10 crc kubenswrapper[4808]: E1002 16:42:10.684185 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:10Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.688845 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.688914 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.688926 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.688955 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.688971 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:10Z","lastTransitionTime":"2025-10-02T16:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:10 crc kubenswrapper[4808]: E1002 16:42:10.706155 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:10Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.711915 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.711975 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.711985 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.712007 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.712021 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:10Z","lastTransitionTime":"2025-10-02T16:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:10 crc kubenswrapper[4808]: E1002 16:42:10.728042 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T16:42:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae2a1887-4119-40ce-b362-d8762c1a02f6\\\",\\\"systemUUID\\\":\\\"36c106c2-22f6-4c16-89bf-0f517cf57f3b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:10Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:10 crc kubenswrapper[4808]: E1002 16:42:10.728346 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.730734 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.730813 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.730838 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.730873 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.730899 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:10Z","lastTransitionTime":"2025-10-02T16:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.834186 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.834273 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.834291 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.834312 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.834328 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:10Z","lastTransitionTime":"2025-10-02T16:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.938014 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.938072 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.938089 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.938114 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:10 crc kubenswrapper[4808]: I1002 16:42:10.938133 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:10Z","lastTransitionTime":"2025-10-02T16:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.041563 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.041641 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.041664 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.041693 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.041712 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:11Z","lastTransitionTime":"2025-10-02T16:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.144737 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.144793 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.144809 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.144831 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.144843 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:11Z","lastTransitionTime":"2025-10-02T16:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.247838 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.247894 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.247907 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.247928 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.247941 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:11Z","lastTransitionTime":"2025-10-02T16:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.350606 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.350648 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.350657 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.350679 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.350693 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:11Z","lastTransitionTime":"2025-10-02T16:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.453844 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.453897 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.453912 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.453934 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.453946 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:11Z","lastTransitionTime":"2025-10-02T16:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.557019 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.557062 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.557076 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.557095 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.557106 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:11Z","lastTransitionTime":"2025-10-02T16:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.660403 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.660462 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.660475 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.660494 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.660504 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:11Z","lastTransitionTime":"2025-10-02T16:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.763855 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.763918 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.763930 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.763954 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.763972 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:11Z","lastTransitionTime":"2025-10-02T16:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.867415 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.867487 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.867509 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.867537 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.867566 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:11Z","lastTransitionTime":"2025-10-02T16:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.970785 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.970845 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.970857 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.970881 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:11 crc kubenswrapper[4808]: I1002 16:42:11.970893 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:11Z","lastTransitionTime":"2025-10-02T16:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.074891 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.074952 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.074969 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.074996 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.075013 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:12Z","lastTransitionTime":"2025-10-02T16:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.178315 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.178383 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.178410 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.178442 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.178466 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:12Z","lastTransitionTime":"2025-10-02T16:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.281935 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.282000 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.282017 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.282042 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.282062 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:12Z","lastTransitionTime":"2025-10-02T16:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.373666 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs\") pod \"network-metrics-daemon-2x552\" (UID: \"e5e40838-d098-402e-b99a-819ce5a5977c\") " pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:12 crc kubenswrapper[4808]: E1002 16:42:12.373851 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 16:42:12 crc kubenswrapper[4808]: E1002 16:42:12.373918 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs podName:e5e40838-d098-402e-b99a-819ce5a5977c nodeName:}" failed. No retries permitted until 2025-10-02 16:43:16.373899393 +0000 UTC m=+163.699428393 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs") pod "network-metrics-daemon-2x552" (UID: "e5e40838-d098-402e-b99a-819ce5a5977c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.384554 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.384591 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.384600 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.384623 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.384636 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:12Z","lastTransitionTime":"2025-10-02T16:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.396136 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.396136 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.396388 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:12 crc kubenswrapper[4808]: E1002 16:42:12.396291 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.396514 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:12 crc kubenswrapper[4808]: E1002 16:42:12.396918 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:12 crc kubenswrapper[4808]: E1002 16:42:12.397397 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:12 crc kubenswrapper[4808]: E1002 16:42:12.397521 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.488119 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.488191 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.488208 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.488261 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.488279 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:12Z","lastTransitionTime":"2025-10-02T16:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.591216 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.591297 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.591307 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.591330 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.591342 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:12Z","lastTransitionTime":"2025-10-02T16:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.694953 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.695013 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.695023 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.695041 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.695054 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:12Z","lastTransitionTime":"2025-10-02T16:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.798398 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.798457 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.798471 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.798490 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.798502 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:12Z","lastTransitionTime":"2025-10-02T16:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.901940 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.901997 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.902008 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.902034 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:12 crc kubenswrapper[4808]: I1002 16:42:12.902045 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:12Z","lastTransitionTime":"2025-10-02T16:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.005392 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.005446 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.005457 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.005475 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.005492 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:13Z","lastTransitionTime":"2025-10-02T16:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.108526 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.108580 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.108596 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.108616 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.108629 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:13Z","lastTransitionTime":"2025-10-02T16:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.211653 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.211721 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.211733 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.211759 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.211772 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:13Z","lastTransitionTime":"2025-10-02T16:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.315117 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.315179 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.315195 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.315224 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.315276 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:13Z","lastTransitionTime":"2025-10-02T16:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.413528 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669c7e8-ad00-4fd5-bcc2-845ffb0ad786\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e8b621e8e173677c37272e90dbe453f2f7b3df6dbb7aac817cb9d846b4cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db3ce94cbdd3f9b78668ebad25e59c58ce105497ae27a488fe0ca7513ec1459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa929245069cf98040d590f5d5936383098e9e4c7273a1ff8368764c1a5f8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f81197e3ee0bf74d6b081ae2f9f8c75c5d6333a0c14ad17c2382949543057baf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c72672ceb6a115555bb9d05d2fbbc5e3b85512412fc31b47cce6c852e386b2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 16:40:36.826996 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 16:40:36.831288 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2586784743/tls.crt::/tmp/serving-cert-2586784743/tls.key\\\\\\\"\\\\nI1002 16:40:52.175988 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 16:40:52.180263 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 16:40:52.180294 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 16:40:52.180319 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 16:40:52.180326 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 16:40:52.184699 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 16:40:52.184731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184736 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 16:40:52.184741 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 16:40:52.184744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 16:40:52.184748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 16:40:52.184750 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 16:40:52.184787 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 16:40:52.188526 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8474ae0a8b5631c5742ce1c0d75311aabdfbcc1979b357f1fbf24f1dfff2af5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47850117c4f6ef546434cd404bcacef82582aa13e0b1223d892cdfe08d174bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.418141 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.418176 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.418185 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.418205 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.418217 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:13Z","lastTransitionTime":"2025-10-02T16:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.425797 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934256ab-6db0-4440-971e-fea9bd9ce459\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe995b6f7e46579fae83eb2918d20e9801ed0f52df4753ad1871c41b9e11abc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a254a3cf9d11b54b681fc4cda92b06f88ff70b87906c439cdaecc58c7b0738f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a254a3cf9d11b54b681fc4cda92b06f88ff70b87906c439cdaecc58c7b0738f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.451138 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5be2f3ad-5544-4d24-9e83-c10214bdb373\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639de8322e0b1203130fa853639da392c845f050cd1d982631c9f5453b563a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://758b452d92fc724952154f775a5821825eb3cd0c9b237b75394df25c5616adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cc44cb0e1e6cc757fb7a5a24f4f10010146a2a0ef2fbc05e552b8bcf11f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af337b752d3e7242919f1bddae39c8c34598e96eca3ea1d0ae1ce603d861ee17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f463f02fb8b7e892102ece1ae89b14f2b452f91de6f7bea1ddfb5f77a9cc612\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a22c13c8327c615bcfc58c49ee75bb14a9d923f41f86164ac61e300c004615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea80b8095f6fc4b4a781913efeaf1f91fbab60556c43aea297b7bdc20a50b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df6da9f987fd49809f7a2abcd10b4d344d2b958f4e1f1c85b3b01be589d6c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.475831 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7d9g4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8871f396-b980-46a8-9477-b3be4889638f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19f7576c6dc4ec19e5895eb22a1790de5928dbd0c4779a000331f3f45df532c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:41:42Z\\\",\\\"message\\\":\\\"2025-10-02T16:40:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_15c55b75-b1a1-40e0-876d-f4ed62e46169\\\\n2025-10-02T16:40:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_15c55b75-b1a1-40e0-876d-f4ed62e46169 to /host/opt/cni/bin/\\\\n2025-10-02T16:40:57Z [verbose] multus-daemon started\\\\n2025-10-02T16:40:57Z [verbose] Readiness Indicator file check\\\\n2025-10-02T16:41:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmcjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7d9g4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.497687 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e441e501-09ee-4e3d-a763-c853de09a02c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T16:42:07Z\\\",\\\"message\\\":\\\"6:42:07.304780 6956 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 16:42:07.305177 6956 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 16:42:07.305193 6956 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 16:42:07.305940 6956 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 16:42:07.305996 6956 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 16:42:07.306026 6956 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 16:42:07.306043 6956 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 16:42:07.306073 6956 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 16:42:07.306102 6956 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 16:42:07.306141 6956 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 16:42:07.306148 6956 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 16:42:07.306165 6956 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 16:42:07.306192 6956 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 16:42:07.306200 6956 factory.go:656] Stopping watch factory\\\\nI1002 16:42:07.306286 6956 ovnkube.go:599] Stopped ovnkube\\\\nI1002 16:42:07.306207 6956 handler.go:208] Removed *v1.Node event handler 7\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T16:42:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zm46w_openshift-ovn-kubernetes(e441e501-09ee-4e3d-a763-c853de09a02c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwkmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zm46w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.514741 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d48974-c7e1-4601-ae3c-e8ecde9e7c79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bea478a450f0237cf2bab3908f785612dc34693aa8dc34ef62548f5c3e60493\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b301868861b72a3ab69aa208069656e3e51e963a7a182d54e3e653642408ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f563406ce5edd396a30e343d195f6fcfef304778c01b126ec3a25ef09f2bc8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7350cbd5de1148646c6c4c4bd6b08dd39ee8619a7122ae1fb4dd4f08be9f00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.520258 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.520287 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.520295 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.520313 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.520326 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:13Z","lastTransitionTime":"2025-10-02T16:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.530874 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bdbb2a7f8349e6d17cbdca70caa08486299cdb25e15667e66e6dbe7e94e706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.550675 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7v5x7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78096007-76e2-47b5-a1c1-549725c4c9a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a9976df508341dd191623e3d35c109dff1fe047aa78908627041eef51887742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tptqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7v5x7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.565031 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f83ce425-101d-4489-94a4-5c256eb29328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8f8be49eb7097ca47982bb71d5cfe07ee96d7f9d44589b9b08cc8507c27c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pw52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7z66r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.578896 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"560139b3-154f-45d4-ae7d-db62351531c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07b9557dcbd29f15e9ae34cc712160a7f5f34330deaa1c08774efd54e68c4ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28b9ee75280e7e9c44e99c1b82c9760990628488cb359247df6483530fa09ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc88fc50e88cd95a703f74122351a86eee2dc44b402e48893690ecd79c9e5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://853d37b9368019b8a47e5b9b00a443319da5e1d3c44c42c2c17ca73da57804f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.592640 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.604361 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.618497 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00e4855926abb52d94061f01474594ce8c41e6cc6de01069febf6068e8649ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e74d1bb2c149ae31a27d8ecf2e870291553e95417cdd0cd1fac5eb39dcbcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.624364 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.624442 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.624452 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.624471 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.624485 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:13Z","lastTransitionTime":"2025-10-02T16:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.632697 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2x552" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e40838-d098-402e-b99a-819ce5a5977c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qz8wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2x552\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.646040 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a0faf439fef3b95a653659ad5c90310ac0cb7ddb6369a9fde4a92abc602599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.660026 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.673200 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ppf7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96534061-3b16-458d-813f-507d5d0acbcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcf5b50942dd32ac0d7803ffed22929a1479ea76387b0b905eed69b915e5af2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8h9sl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ppf7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.694312 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-75l8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5716183c-6301-4312-923c-e34254575a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7547c65d0347083301861e105e5190dce01f2109c8ce2ca17dfaa7a15dfd7219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e06dc53b28d9fe76c6a48af4794c844b0ff7932bc3e0532efe477f93f1f13b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d795991249cf9439a9cd293cd9f5cb68b18c1c2743ed5d667f4da7cbd95bb51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05da11424cf25890377b88d0cbcda571a8b838991b79924a4162c083d6e70d57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7869302029a61438ed856a1f80d8554191cf8af789ed2eff496a063d152a75c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://437833193f8c934d59411bbb47bbc5e6f3ea028e7848df59d3658261cbae8b95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6193445636e6c02ec8500237acf951cde44ff9fb18fc66166f6e64a621836cdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T16:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T16:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qqqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-75l8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.713543 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4021608b-add3-4b37-b523-f1d3bf65ca05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T16:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f7426ac250fc0c8a7afe7b809d6ba62dc69ae4ff0f2e595027f86958d505ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ef5f79f90831cb32131db7ea8d2f305c3664bd9d06dd98c007b74378f1bf1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T16:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vd76q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T16:41:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llclq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T16:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.727794 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.727837 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.727846 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.727862 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.727871 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:13Z","lastTransitionTime":"2025-10-02T16:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.831548 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.831608 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.831620 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.831643 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.831654 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:13Z","lastTransitionTime":"2025-10-02T16:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.934273 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.934313 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.934321 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.934339 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:13 crc kubenswrapper[4808]: I1002 16:42:13.934350 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:13Z","lastTransitionTime":"2025-10-02T16:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.036929 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.036971 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.036979 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.036997 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.037008 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:14Z","lastTransitionTime":"2025-10-02T16:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.140354 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.140401 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.140410 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.140427 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.140437 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:14Z","lastTransitionTime":"2025-10-02T16:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.243969 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.244038 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.244051 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.244417 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.244456 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:14Z","lastTransitionTime":"2025-10-02T16:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.347486 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.347546 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.347558 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.347581 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.347594 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:14Z","lastTransitionTime":"2025-10-02T16:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.395537 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.395582 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.395652 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:14 crc kubenswrapper[4808]: E1002 16:42:14.395716 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.395555 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:14 crc kubenswrapper[4808]: E1002 16:42:14.395854 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:14 crc kubenswrapper[4808]: E1002 16:42:14.396006 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:14 crc kubenswrapper[4808]: E1002 16:42:14.396104 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.450877 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.450952 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.450965 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.450988 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.451006 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:14Z","lastTransitionTime":"2025-10-02T16:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.554132 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.554185 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.554206 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.554229 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.554265 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:14Z","lastTransitionTime":"2025-10-02T16:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.657703 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.657751 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.657762 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.657781 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.657794 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:14Z","lastTransitionTime":"2025-10-02T16:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.761260 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.761329 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.761348 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.761368 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.761381 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:14Z","lastTransitionTime":"2025-10-02T16:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.864323 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.864362 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.864372 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.864390 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.864402 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:14Z","lastTransitionTime":"2025-10-02T16:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.967273 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.967330 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.967343 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.967370 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:14 crc kubenswrapper[4808]: I1002 16:42:14.967385 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:14Z","lastTransitionTime":"2025-10-02T16:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.070804 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.070863 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.070877 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.070922 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.070935 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:15Z","lastTransitionTime":"2025-10-02T16:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.174007 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.174064 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.174078 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.174100 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.174118 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:15Z","lastTransitionTime":"2025-10-02T16:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.277163 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.277209 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.277218 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.277255 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.277267 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:15Z","lastTransitionTime":"2025-10-02T16:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.379682 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.379727 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.379736 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.379752 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.379762 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:15Z","lastTransitionTime":"2025-10-02T16:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.483627 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.483672 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.483682 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.483700 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.483712 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:15Z","lastTransitionTime":"2025-10-02T16:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.586875 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.586935 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.586948 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.586967 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.586981 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:15Z","lastTransitionTime":"2025-10-02T16:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.690293 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.690364 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.690375 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.690399 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.690413 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:15Z","lastTransitionTime":"2025-10-02T16:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.793552 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.793631 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.793658 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.793685 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.793705 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:15Z","lastTransitionTime":"2025-10-02T16:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.896976 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.897034 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.897053 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.897080 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:15 crc kubenswrapper[4808]: I1002 16:42:15.897099 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:15Z","lastTransitionTime":"2025-10-02T16:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.000766 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.000823 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.000841 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.000865 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.000884 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:16Z","lastTransitionTime":"2025-10-02T16:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.104346 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.104414 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.104430 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.104452 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.104469 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:16Z","lastTransitionTime":"2025-10-02T16:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.207750 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.207800 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.207812 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.207835 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.207853 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:16Z","lastTransitionTime":"2025-10-02T16:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.311573 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.311642 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.311658 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.311682 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.311692 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:16Z","lastTransitionTime":"2025-10-02T16:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.395481 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.395617 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:16 crc kubenswrapper[4808]: E1002 16:42:16.395684 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.395482 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.395516 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:16 crc kubenswrapper[4808]: E1002 16:42:16.395821 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:16 crc kubenswrapper[4808]: E1002 16:42:16.395904 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:16 crc kubenswrapper[4808]: E1002 16:42:16.396024 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.414613 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.414653 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.414662 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.414680 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.414695 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:16Z","lastTransitionTime":"2025-10-02T16:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.518008 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.518050 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.518060 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.518077 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.518088 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:16Z","lastTransitionTime":"2025-10-02T16:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.620985 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.621080 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.621107 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.621140 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.621165 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:16Z","lastTransitionTime":"2025-10-02T16:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.724488 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.724560 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.724579 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.724613 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.724633 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:16Z","lastTransitionTime":"2025-10-02T16:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.827790 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.827845 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.827860 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.827883 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.827900 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:16Z","lastTransitionTime":"2025-10-02T16:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.931350 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.931405 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.931420 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.931441 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:16 crc kubenswrapper[4808]: I1002 16:42:16.931458 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:16Z","lastTransitionTime":"2025-10-02T16:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.035348 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.035417 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.035435 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.035462 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.035485 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:17Z","lastTransitionTime":"2025-10-02T16:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.138588 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.138656 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.138667 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.138694 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.138710 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:17Z","lastTransitionTime":"2025-10-02T16:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.241584 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.241635 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.241645 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.241664 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.241677 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:17Z","lastTransitionTime":"2025-10-02T16:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.344467 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.344546 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.344561 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.344582 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.344622 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:17Z","lastTransitionTime":"2025-10-02T16:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.448042 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.448105 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.448120 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.448144 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.448173 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:17Z","lastTransitionTime":"2025-10-02T16:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.551389 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.551431 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.551441 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.551458 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.551467 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:17Z","lastTransitionTime":"2025-10-02T16:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.653784 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.653864 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.653887 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.653919 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.653946 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:17Z","lastTransitionTime":"2025-10-02T16:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.756476 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.756515 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.756527 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.756546 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.756560 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:17Z","lastTransitionTime":"2025-10-02T16:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.859360 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.859398 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.859406 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.859426 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.859438 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:17Z","lastTransitionTime":"2025-10-02T16:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.962939 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.962979 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.962991 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.963017 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:17 crc kubenswrapper[4808]: I1002 16:42:17.963030 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:17Z","lastTransitionTime":"2025-10-02T16:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.065799 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.065854 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.065865 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.065885 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.065900 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:18Z","lastTransitionTime":"2025-10-02T16:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.168693 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.168734 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.168743 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.168760 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.168770 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:18Z","lastTransitionTime":"2025-10-02T16:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.272081 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.272125 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.272134 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.272152 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.272164 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:18Z","lastTransitionTime":"2025-10-02T16:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.374746 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.374792 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.374801 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.374819 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.374830 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:18Z","lastTransitionTime":"2025-10-02T16:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.395453 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.395498 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.395526 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.395597 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:18 crc kubenswrapper[4808]: E1002 16:42:18.395597 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:18 crc kubenswrapper[4808]: E1002 16:42:18.395685 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:18 crc kubenswrapper[4808]: E1002 16:42:18.395854 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:18 crc kubenswrapper[4808]: E1002 16:42:18.395946 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.478389 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.478449 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.478461 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.478515 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.478529 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:18Z","lastTransitionTime":"2025-10-02T16:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.581998 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.582080 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.582102 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.582139 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.582158 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:18Z","lastTransitionTime":"2025-10-02T16:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.685611 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.685697 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.685721 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.685752 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.685779 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:18Z","lastTransitionTime":"2025-10-02T16:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.789420 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.789486 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.789497 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.789513 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.789525 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:18Z","lastTransitionTime":"2025-10-02T16:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.893953 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.894088 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.894115 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.894141 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.894163 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:18Z","lastTransitionTime":"2025-10-02T16:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.999368 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.999478 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.999496 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:18 crc kubenswrapper[4808]: I1002 16:42:18.999521 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:18.999535 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:18Z","lastTransitionTime":"2025-10-02T16:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.103156 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.103219 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.103278 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.103309 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.103326 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:19Z","lastTransitionTime":"2025-10-02T16:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.207309 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.207651 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.207797 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.207876 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.207961 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:19Z","lastTransitionTime":"2025-10-02T16:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.311471 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.311541 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.311563 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.311596 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.311899 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:19Z","lastTransitionTime":"2025-10-02T16:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.415841 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.415926 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.415946 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.415978 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.416005 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:19Z","lastTransitionTime":"2025-10-02T16:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.519118 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.519229 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.519280 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.519307 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.519328 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:19Z","lastTransitionTime":"2025-10-02T16:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.621704 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.621766 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.621785 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.621810 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.621827 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:19Z","lastTransitionTime":"2025-10-02T16:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.724993 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.725133 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.725156 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.725191 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.725211 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:19Z","lastTransitionTime":"2025-10-02T16:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.828559 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.828634 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.828655 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.828682 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.828702 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:19Z","lastTransitionTime":"2025-10-02T16:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.932902 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.932962 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.932971 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.932995 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:19 crc kubenswrapper[4808]: I1002 16:42:19.933006 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:19Z","lastTransitionTime":"2025-10-02T16:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.036763 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.036820 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.036837 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.036862 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.036880 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:20Z","lastTransitionTime":"2025-10-02T16:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.140738 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.140792 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.140805 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.140824 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.140836 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:20Z","lastTransitionTime":"2025-10-02T16:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.245551 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.245621 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.245639 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.245668 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.245692 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:20Z","lastTransitionTime":"2025-10-02T16:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.349425 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.349488 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.349500 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.349522 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.349536 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:20Z","lastTransitionTime":"2025-10-02T16:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.395433 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.395448 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.395574 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:20 crc kubenswrapper[4808]: E1002 16:42:20.396083 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.395543 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:20 crc kubenswrapper[4808]: E1002 16:42:20.396405 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:20 crc kubenswrapper[4808]: E1002 16:42:20.396433 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:20 crc kubenswrapper[4808]: E1002 16:42:20.397060 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.453424 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.453821 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.453963 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.454076 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.454175 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:20Z","lastTransitionTime":"2025-10-02T16:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.557438 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.557487 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.557501 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.557523 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.557538 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:20Z","lastTransitionTime":"2025-10-02T16:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.660831 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.660882 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.660891 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.660952 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.660965 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:20Z","lastTransitionTime":"2025-10-02T16:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.764407 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.764486 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.764505 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.764533 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.764557 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:20Z","lastTransitionTime":"2025-10-02T16:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.867954 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.868030 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.868049 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.868072 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.868089 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:20Z","lastTransitionTime":"2025-10-02T16:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.970723 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.970781 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.970797 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.970818 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:20 crc kubenswrapper[4808]: I1002 16:42:20.970832 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:20Z","lastTransitionTime":"2025-10-02T16:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.074380 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.074465 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.074488 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.074521 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.074540 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:21Z","lastTransitionTime":"2025-10-02T16:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.106905 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.106956 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.106975 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.107002 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.107021 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T16:42:21Z","lastTransitionTime":"2025-10-02T16:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.170844 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77"] Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.172159 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.178937 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.179915 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.180014 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.181697 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.193627 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-ppf7n" podStartSLOduration=87.193603809 podStartE2EDuration="1m27.193603809s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:42:21.192976143 +0000 UTC m=+108.518505163" watchObservedRunningTime="2025-10-02 16:42:21.193603809 +0000 UTC m=+108.519132819" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.235530 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-75l8m" podStartSLOduration=87.235501851 podStartE2EDuration="1m27.235501851s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:42:21.216180686 +0000 UTC m=+108.541709746" watchObservedRunningTime="2025-10-02 16:42:21.235501851 +0000 UTC m=+108.561030851" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.235743 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llclq" podStartSLOduration=87.235739507 podStartE2EDuration="1m27.235739507s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:42:21.235330877 +0000 UTC m=+108.560859887" watchObservedRunningTime="2025-10-02 16:42:21.235739507 +0000 UTC m=+108.561268507" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.284968 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c34457da-c246-461a-ae09-1129236ceee0-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-k7l77\" (UID: \"c34457da-c246-461a-ae09-1129236ceee0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.285049 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c34457da-c246-461a-ae09-1129236ceee0-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-k7l77\" (UID: \"c34457da-c246-461a-ae09-1129236ceee0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.285069 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c34457da-c246-461a-ae09-1129236ceee0-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-k7l77\" (UID: \"c34457da-c246-461a-ae09-1129236ceee0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.285100 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c34457da-c246-461a-ae09-1129236ceee0-service-ca\") pod \"cluster-version-operator-5c965bbfc6-k7l77\" (UID: \"c34457da-c246-461a-ae09-1129236ceee0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.285194 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c34457da-c246-461a-ae09-1129236ceee0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-k7l77\" (UID: \"c34457da-c246-461a-ae09-1129236ceee0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.300785 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=89.300764601 podStartE2EDuration="1m29.300764601s" podCreationTimestamp="2025-10-02 16:40:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:42:21.298760171 +0000 UTC m=+108.624289171" watchObservedRunningTime="2025-10-02 16:42:21.300764601 +0000 UTC m=+108.626293601" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.315386 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-7d9g4" podStartSLOduration=87.315357028 podStartE2EDuration="1m27.315357028s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:42:21.314981499 +0000 UTC m=+108.640510499" watchObservedRunningTime="2025-10-02 16:42:21.315357028 +0000 UTC m=+108.640886038" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.371704 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=89.371682993 podStartE2EDuration="1m29.371682993s" podCreationTimestamp="2025-10-02 16:40:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:42:21.371038007 +0000 UTC m=+108.696567007" watchObservedRunningTime="2025-10-02 16:42:21.371682993 +0000 UTC m=+108.697212003" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.385788 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c34457da-c246-461a-ae09-1129236ceee0-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-k7l77\" (UID: \"c34457da-c246-461a-ae09-1129236ceee0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.385850 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c34457da-c246-461a-ae09-1129236ceee0-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-k7l77\" (UID: \"c34457da-c246-461a-ae09-1129236ceee0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.385874 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c34457da-c246-461a-ae09-1129236ceee0-service-ca\") pod \"cluster-version-operator-5c965bbfc6-k7l77\" (UID: \"c34457da-c246-461a-ae09-1129236ceee0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.385875 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c34457da-c246-461a-ae09-1129236ceee0-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-k7l77\" (UID: \"c34457da-c246-461a-ae09-1129236ceee0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.385944 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c34457da-c246-461a-ae09-1129236ceee0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-k7l77\" (UID: \"c34457da-c246-461a-ae09-1129236ceee0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.385990 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c34457da-c246-461a-ae09-1129236ceee0-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-k7l77\" (UID: \"c34457da-c246-461a-ae09-1129236ceee0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.386369 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c34457da-c246-461a-ae09-1129236ceee0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-k7l77\" (UID: \"c34457da-c246-461a-ae09-1129236ceee0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.387554 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c34457da-c246-461a-ae09-1129236ceee0-service-ca\") pod \"cluster-version-operator-5c965bbfc6-k7l77\" (UID: \"c34457da-c246-461a-ae09-1129236ceee0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.395127 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c34457da-c246-461a-ae09-1129236ceee0-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-k7l77\" (UID: \"c34457da-c246-461a-ae09-1129236ceee0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.400518 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-7v5x7" podStartSLOduration=89.400500428 podStartE2EDuration="1m29.400500428s" podCreationTimestamp="2025-10-02 16:40:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:42:21.399821951 +0000 UTC m=+108.725350971" watchObservedRunningTime="2025-10-02 16:42:21.400500428 +0000 UTC m=+108.726029428" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.400631 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=26.400627401 podStartE2EDuration="26.400627401s" podCreationTimestamp="2025-10-02 16:41:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:42:21.385554152 +0000 UTC m=+108.711083152" watchObservedRunningTime="2025-10-02 16:42:21.400627401 +0000 UTC m=+108.726156401" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.403610 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c34457da-c246-461a-ae09-1129236ceee0-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-k7l77\" (UID: \"c34457da-c246-461a-ae09-1129236ceee0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.427847 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podStartSLOduration=87.427820484 podStartE2EDuration="1m27.427820484s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:42:21.414054758 +0000 UTC m=+108.739583758" watchObservedRunningTime="2025-10-02 16:42:21.427820484 +0000 UTC m=+108.753349484" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.443067 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=85.443039487 podStartE2EDuration="1m25.443039487s" podCreationTimestamp="2025-10-02 16:40:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:42:21.441452957 +0000 UTC m=+108.766981967" watchObservedRunningTime="2025-10-02 16:42:21.443039487 +0000 UTC m=+108.768568507" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.499881 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77" Oct 02 16:42:21 crc kubenswrapper[4808]: I1002 16:42:21.523961 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=52.523927649 podStartE2EDuration="52.523927649s" podCreationTimestamp="2025-10-02 16:41:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:42:21.52158994 +0000 UTC m=+108.847118950" watchObservedRunningTime="2025-10-02 16:42:21.523927649 +0000 UTC m=+108.849456649" Oct 02 16:42:22 crc kubenswrapper[4808]: I1002 16:42:22.028706 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77" event={"ID":"c34457da-c246-461a-ae09-1129236ceee0","Type":"ContainerStarted","Data":"073e2742abc052e63dd095d38f7084f61914456e37b1c798f9d4b99e484c810d"} Oct 02 16:42:22 crc kubenswrapper[4808]: I1002 16:42:22.028792 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77" event={"ID":"c34457da-c246-461a-ae09-1129236ceee0","Type":"ContainerStarted","Data":"7484aea6e316d87befb97a4328efacb12ed2d2bbfafbecf1e83db10a34abf8eb"} Oct 02 16:42:22 crc kubenswrapper[4808]: I1002 16:42:22.050924 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k7l77" podStartSLOduration=88.05089756 podStartE2EDuration="1m28.05089756s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:42:22.050065159 +0000 UTC m=+109.375594159" watchObservedRunningTime="2025-10-02 16:42:22.05089756 +0000 UTC m=+109.376426600" Oct 02 16:42:22 crc kubenswrapper[4808]: I1002 16:42:22.395639 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:22 crc kubenswrapper[4808]: I1002 16:42:22.395684 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:22 crc kubenswrapper[4808]: I1002 16:42:22.395652 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:22 crc kubenswrapper[4808]: E1002 16:42:22.395798 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:22 crc kubenswrapper[4808]: I1002 16:42:22.395901 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:22 crc kubenswrapper[4808]: E1002 16:42:22.396021 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:22 crc kubenswrapper[4808]: E1002 16:42:22.396146 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:22 crc kubenswrapper[4808]: E1002 16:42:22.396222 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:23 crc kubenswrapper[4808]: I1002 16:42:23.399344 4808 scope.go:117] "RemoveContainer" containerID="f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97" Oct 02 16:42:23 crc kubenswrapper[4808]: E1002 16:42:23.399675 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zm46w_openshift-ovn-kubernetes(e441e501-09ee-4e3d-a763-c853de09a02c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" Oct 02 16:42:24 crc kubenswrapper[4808]: I1002 16:42:24.395320 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:24 crc kubenswrapper[4808]: I1002 16:42:24.395378 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:24 crc kubenswrapper[4808]: E1002 16:42:24.395520 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:24 crc kubenswrapper[4808]: E1002 16:42:24.395682 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:24 crc kubenswrapper[4808]: I1002 16:42:24.395850 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:24 crc kubenswrapper[4808]: E1002 16:42:24.395996 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:24 crc kubenswrapper[4808]: I1002 16:42:24.396092 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:24 crc kubenswrapper[4808]: E1002 16:42:24.396211 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:26 crc kubenswrapper[4808]: I1002 16:42:26.395293 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:26 crc kubenswrapper[4808]: E1002 16:42:26.395525 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:26 crc kubenswrapper[4808]: I1002 16:42:26.395834 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:26 crc kubenswrapper[4808]: I1002 16:42:26.395862 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:26 crc kubenswrapper[4808]: I1002 16:42:26.396058 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:26 crc kubenswrapper[4808]: E1002 16:42:26.396217 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:26 crc kubenswrapper[4808]: E1002 16:42:26.396045 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:26 crc kubenswrapper[4808]: E1002 16:42:26.396406 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:28 crc kubenswrapper[4808]: I1002 16:42:28.395378 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:28 crc kubenswrapper[4808]: I1002 16:42:28.395471 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:28 crc kubenswrapper[4808]: E1002 16:42:28.395593 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:28 crc kubenswrapper[4808]: I1002 16:42:28.395483 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:28 crc kubenswrapper[4808]: E1002 16:42:28.395684 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:28 crc kubenswrapper[4808]: I1002 16:42:28.395395 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:28 crc kubenswrapper[4808]: E1002 16:42:28.395886 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:28 crc kubenswrapper[4808]: E1002 16:42:28.395976 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:29 crc kubenswrapper[4808]: I1002 16:42:29.060848 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7d9g4_8871f396-b980-46a8-9477-b3be4889638f/kube-multus/1.log" Oct 02 16:42:29 crc kubenswrapper[4808]: I1002 16:42:29.061663 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7d9g4_8871f396-b980-46a8-9477-b3be4889638f/kube-multus/0.log" Oct 02 16:42:29 crc kubenswrapper[4808]: I1002 16:42:29.061763 4808 generic.go:334] "Generic (PLEG): container finished" podID="8871f396-b980-46a8-9477-b3be4889638f" containerID="19f7576c6dc4ec19e5895eb22a1790de5928dbd0c4779a000331f3f45df532c1" exitCode=1 Oct 02 16:42:29 crc kubenswrapper[4808]: I1002 16:42:29.061825 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7d9g4" event={"ID":"8871f396-b980-46a8-9477-b3be4889638f","Type":"ContainerDied","Data":"19f7576c6dc4ec19e5895eb22a1790de5928dbd0c4779a000331f3f45df532c1"} Oct 02 16:42:29 crc kubenswrapper[4808]: I1002 16:42:29.061895 4808 scope.go:117] "RemoveContainer" containerID="ee7df3d0fa22a99dde956b09456b6848774d5391e2374a978ad1c9fa9fc6c053" Oct 02 16:42:29 crc kubenswrapper[4808]: I1002 16:42:29.062525 4808 scope.go:117] "RemoveContainer" containerID="19f7576c6dc4ec19e5895eb22a1790de5928dbd0c4779a000331f3f45df532c1" Oct 02 16:42:29 crc kubenswrapper[4808]: E1002 16:42:29.062926 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-7d9g4_openshift-multus(8871f396-b980-46a8-9477-b3be4889638f)\"" pod="openshift-multus/multus-7d9g4" podUID="8871f396-b980-46a8-9477-b3be4889638f" Oct 02 16:42:30 crc kubenswrapper[4808]: I1002 16:42:30.069392 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7d9g4_8871f396-b980-46a8-9477-b3be4889638f/kube-multus/1.log" Oct 02 16:42:30 crc kubenswrapper[4808]: I1002 16:42:30.395675 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:30 crc kubenswrapper[4808]: I1002 16:42:30.395827 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:30 crc kubenswrapper[4808]: I1002 16:42:30.396585 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:30 crc kubenswrapper[4808]: I1002 16:42:30.396833 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:30 crc kubenswrapper[4808]: E1002 16:42:30.397429 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:30 crc kubenswrapper[4808]: E1002 16:42:30.397441 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:30 crc kubenswrapper[4808]: E1002 16:42:30.397439 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:30 crc kubenswrapper[4808]: E1002 16:42:30.397565 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:32 crc kubenswrapper[4808]: I1002 16:42:32.395560 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:32 crc kubenswrapper[4808]: I1002 16:42:32.395639 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:32 crc kubenswrapper[4808]: E1002 16:42:32.395754 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:32 crc kubenswrapper[4808]: I1002 16:42:32.395573 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:32 crc kubenswrapper[4808]: E1002 16:42:32.395909 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:32 crc kubenswrapper[4808]: E1002 16:42:32.396085 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:32 crc kubenswrapper[4808]: I1002 16:42:32.396167 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:32 crc kubenswrapper[4808]: E1002 16:42:32.396714 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:33 crc kubenswrapper[4808]: E1002 16:42:33.322183 4808 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 02 16:42:33 crc kubenswrapper[4808]: E1002 16:42:33.510615 4808 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 16:42:34 crc kubenswrapper[4808]: I1002 16:42:34.395968 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:34 crc kubenswrapper[4808]: I1002 16:42:34.396032 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:34 crc kubenswrapper[4808]: I1002 16:42:34.396115 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:34 crc kubenswrapper[4808]: E1002 16:42:34.396312 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:34 crc kubenswrapper[4808]: I1002 16:42:34.396417 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:34 crc kubenswrapper[4808]: E1002 16:42:34.396755 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:34 crc kubenswrapper[4808]: E1002 16:42:34.396904 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:34 crc kubenswrapper[4808]: E1002 16:42:34.396957 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:36 crc kubenswrapper[4808]: I1002 16:42:36.395699 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:36 crc kubenswrapper[4808]: I1002 16:42:36.395749 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:36 crc kubenswrapper[4808]: I1002 16:42:36.395770 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:36 crc kubenswrapper[4808]: I1002 16:42:36.395716 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:36 crc kubenswrapper[4808]: E1002 16:42:36.395885 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:36 crc kubenswrapper[4808]: E1002 16:42:36.396402 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:36 crc kubenswrapper[4808]: E1002 16:42:36.397149 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:36 crc kubenswrapper[4808]: E1002 16:42:36.397267 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:38 crc kubenswrapper[4808]: I1002 16:42:38.395415 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:38 crc kubenswrapper[4808]: I1002 16:42:38.395498 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:38 crc kubenswrapper[4808]: E1002 16:42:38.395585 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:38 crc kubenswrapper[4808]: I1002 16:42:38.395611 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:38 crc kubenswrapper[4808]: I1002 16:42:38.395874 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:38 crc kubenswrapper[4808]: E1002 16:42:38.395847 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:38 crc kubenswrapper[4808]: E1002 16:42:38.396672 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:38 crc kubenswrapper[4808]: E1002 16:42:38.396763 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:38 crc kubenswrapper[4808]: I1002 16:42:38.397212 4808 scope.go:117] "RemoveContainer" containerID="f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97" Oct 02 16:42:38 crc kubenswrapper[4808]: E1002 16:42:38.398336 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zm46w_openshift-ovn-kubernetes(e441e501-09ee-4e3d-a763-c853de09a02c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" Oct 02 16:42:38 crc kubenswrapper[4808]: E1002 16:42:38.512596 4808 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 16:42:40 crc kubenswrapper[4808]: I1002 16:42:40.395747 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:40 crc kubenswrapper[4808]: I1002 16:42:40.395747 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:40 crc kubenswrapper[4808]: I1002 16:42:40.396457 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:40 crc kubenswrapper[4808]: E1002 16:42:40.396681 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:40 crc kubenswrapper[4808]: I1002 16:42:40.396767 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:40 crc kubenswrapper[4808]: E1002 16:42:40.396845 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:40 crc kubenswrapper[4808]: E1002 16:42:40.397007 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:40 crc kubenswrapper[4808]: E1002 16:42:40.397159 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:42 crc kubenswrapper[4808]: I1002 16:42:42.395158 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:42 crc kubenswrapper[4808]: I1002 16:42:42.395290 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:42 crc kubenswrapper[4808]: I1002 16:42:42.395370 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:42 crc kubenswrapper[4808]: I1002 16:42:42.396337 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:42 crc kubenswrapper[4808]: E1002 16:42:42.396633 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:42 crc kubenswrapper[4808]: E1002 16:42:42.396697 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:42 crc kubenswrapper[4808]: E1002 16:42:42.396779 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:42 crc kubenswrapper[4808]: E1002 16:42:42.396844 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:43 crc kubenswrapper[4808]: E1002 16:42:43.513353 4808 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 16:42:44 crc kubenswrapper[4808]: I1002 16:42:44.395664 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:44 crc kubenswrapper[4808]: I1002 16:42:44.395745 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:44 crc kubenswrapper[4808]: I1002 16:42:44.395687 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:44 crc kubenswrapper[4808]: I1002 16:42:44.395779 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:44 crc kubenswrapper[4808]: E1002 16:42:44.396175 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:44 crc kubenswrapper[4808]: I1002 16:42:44.396267 4808 scope.go:117] "RemoveContainer" containerID="19f7576c6dc4ec19e5895eb22a1790de5928dbd0c4779a000331f3f45df532c1" Oct 02 16:42:44 crc kubenswrapper[4808]: E1002 16:42:44.396209 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:44 crc kubenswrapper[4808]: E1002 16:42:44.396319 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:44 crc kubenswrapper[4808]: E1002 16:42:44.396617 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:45 crc kubenswrapper[4808]: I1002 16:42:45.130788 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7d9g4_8871f396-b980-46a8-9477-b3be4889638f/kube-multus/1.log" Oct 02 16:42:45 crc kubenswrapper[4808]: I1002 16:42:45.131212 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7d9g4" event={"ID":"8871f396-b980-46a8-9477-b3be4889638f","Type":"ContainerStarted","Data":"6e7bed9dd604e3e5fa64c1b2ee8dba75032ba4b53407915796d9780631b4ec78"} Oct 02 16:42:46 crc kubenswrapper[4808]: I1002 16:42:46.395318 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:46 crc kubenswrapper[4808]: I1002 16:42:46.395424 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:46 crc kubenswrapper[4808]: E1002 16:42:46.396202 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:46 crc kubenswrapper[4808]: I1002 16:42:46.395481 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:46 crc kubenswrapper[4808]: I1002 16:42:46.395449 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:46 crc kubenswrapper[4808]: E1002 16:42:46.396346 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:46 crc kubenswrapper[4808]: E1002 16:42:46.396507 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:46 crc kubenswrapper[4808]: E1002 16:42:46.396636 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:48 crc kubenswrapper[4808]: I1002 16:42:48.395603 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:48 crc kubenswrapper[4808]: I1002 16:42:48.395688 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:48 crc kubenswrapper[4808]: I1002 16:42:48.395745 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:48 crc kubenswrapper[4808]: I1002 16:42:48.395755 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:48 crc kubenswrapper[4808]: E1002 16:42:48.395791 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:48 crc kubenswrapper[4808]: E1002 16:42:48.396162 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:48 crc kubenswrapper[4808]: E1002 16:42:48.395986 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:48 crc kubenswrapper[4808]: E1002 16:42:48.396499 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:48 crc kubenswrapper[4808]: E1002 16:42:48.515600 4808 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 16:42:49 crc kubenswrapper[4808]: I1002 16:42:49.397396 4808 scope.go:117] "RemoveContainer" containerID="f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97" Oct 02 16:42:50 crc kubenswrapper[4808]: I1002 16:42:50.157115 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zm46w_e441e501-09ee-4e3d-a763-c853de09a02c/ovnkube-controller/3.log" Oct 02 16:42:50 crc kubenswrapper[4808]: I1002 16:42:50.160494 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerStarted","Data":"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419"} Oct 02 16:42:50 crc kubenswrapper[4808]: I1002 16:42:50.161097 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:42:50 crc kubenswrapper[4808]: I1002 16:42:50.197578 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" podStartSLOduration=116.197553367 podStartE2EDuration="1m56.197553367s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:42:50.196368377 +0000 UTC m=+137.521897387" watchObservedRunningTime="2025-10-02 16:42:50.197553367 +0000 UTC m=+137.523082367" Oct 02 16:42:50 crc kubenswrapper[4808]: I1002 16:42:50.393699 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-2x552"] Oct 02 16:42:50 crc kubenswrapper[4808]: I1002 16:42:50.393911 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:50 crc kubenswrapper[4808]: E1002 16:42:50.394050 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:50 crc kubenswrapper[4808]: I1002 16:42:50.395903 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:50 crc kubenswrapper[4808]: E1002 16:42:50.396163 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:50 crc kubenswrapper[4808]: I1002 16:42:50.396351 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:50 crc kubenswrapper[4808]: E1002 16:42:50.396500 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:50 crc kubenswrapper[4808]: I1002 16:42:50.396628 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:50 crc kubenswrapper[4808]: E1002 16:42:50.396761 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:52 crc kubenswrapper[4808]: I1002 16:42:52.395645 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:52 crc kubenswrapper[4808]: I1002 16:42:52.395776 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:52 crc kubenswrapper[4808]: I1002 16:42:52.395666 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:52 crc kubenswrapper[4808]: E1002 16:42:52.395876 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 16:42:52 crc kubenswrapper[4808]: E1002 16:42:52.395987 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 16:42:52 crc kubenswrapper[4808]: E1002 16:42:52.396135 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2x552" podUID="e5e40838-d098-402e-b99a-819ce5a5977c" Oct 02 16:42:52 crc kubenswrapper[4808]: I1002 16:42:52.396220 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:52 crc kubenswrapper[4808]: E1002 16:42:52.396320 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 16:42:54 crc kubenswrapper[4808]: I1002 16:42:54.395330 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:42:54 crc kubenswrapper[4808]: I1002 16:42:54.395387 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:42:54 crc kubenswrapper[4808]: I1002 16:42:54.395389 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:42:54 crc kubenswrapper[4808]: I1002 16:42:54.395323 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:42:54 crc kubenswrapper[4808]: I1002 16:42:54.398925 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 02 16:42:54 crc kubenswrapper[4808]: I1002 16:42:54.398935 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 02 16:42:54 crc kubenswrapper[4808]: I1002 16:42:54.398980 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 02 16:42:54 crc kubenswrapper[4808]: I1002 16:42:54.398930 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 02 16:42:54 crc kubenswrapper[4808]: I1002 16:42:54.399129 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 02 16:42:54 crc kubenswrapper[4808]: I1002 16:42:54.402868 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 02 16:43:00 crc kubenswrapper[4808]: I1002 16:43:00.180910 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:00 crc kubenswrapper[4808]: E1002 16:43:00.181450 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:45:02.181409635 +0000 UTC m=+269.506938675 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:00 crc kubenswrapper[4808]: I1002 16:43:00.181644 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:43:00 crc kubenswrapper[4808]: I1002 16:43:00.181768 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:43:00 crc kubenswrapper[4808]: I1002 16:43:00.183027 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:43:00 crc kubenswrapper[4808]: I1002 16:43:00.190922 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:43:00 crc kubenswrapper[4808]: I1002 16:43:00.282649 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:43:00 crc kubenswrapper[4808]: I1002 16:43:00.282789 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:43:00 crc kubenswrapper[4808]: I1002 16:43:00.287284 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:43:00 crc kubenswrapper[4808]: I1002 16:43:00.288590 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:43:00 crc kubenswrapper[4808]: I1002 16:43:00.412093 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:43:00 crc kubenswrapper[4808]: I1002 16:43:00.422872 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 16:43:00 crc kubenswrapper[4808]: I1002 16:43:00.446823 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 16:43:00 crc kubenswrapper[4808]: W1002 16:43:00.910648 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-f053eb2a4ca73502e7fca6cf954e966fae7d12fef4fa10d2ea49f74996b27248 WatchSource:0}: Error finding container f053eb2a4ca73502e7fca6cf954e966fae7d12fef4fa10d2ea49f74996b27248: Status 404 returned error can't find the container with id f053eb2a4ca73502e7fca6cf954e966fae7d12fef4fa10d2ea49f74996b27248 Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.205373 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"9213ecaa6bc52061abb3b94e41cbf5f71b2171d508888dc1b2e9eb263ee1b956"} Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.214616 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f053eb2a4ca73502e7fca6cf954e966fae7d12fef4fa10d2ea49f74996b27248"} Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.218330 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"8044b065d082cec8a9ab5b0ef775c283b0cb085f70c3da38d46f54e0ccbf229b"} Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.218411 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e1d921a68d83404f53dea683119338dc4337215b1e603f94f9a68ac4442611de"} Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.848805 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.907888 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zqnpj"] Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.908927 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ffv8w"] Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.909461 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ffv8w" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.910143 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-zqnpj" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.928975 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-lkxxm"] Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.929620 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-lkxxm" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.930276 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m"] Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.931023 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.931051 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.932889 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5zh9"] Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.933396 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5zh9" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.933503 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.933928 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.934036 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.934155 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.934334 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.934564 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.934770 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.934925 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.935047 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.935211 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.936695 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.937204 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.937544 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.937655 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-dlddx"] Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.937936 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.937994 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2bgkb"] Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.938161 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.938337 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.938427 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.938475 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4dk9g"] Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.938630 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.938742 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4dk9g" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.939070 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.939896 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.940458 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.937954 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.940512 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.941385 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.944518 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8974h"] Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.945360 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9xvr"] Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.945392 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.946496 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9xvr" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.946622 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-t4qrz"] Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.947154 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-t4qrz" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.953911 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-qsqkp"] Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.954376 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-lbw7q"] Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.954787 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf"] Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.955314 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.955974 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-qsqkp" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.956248 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lbw7q" Oct 02 16:43:01 crc kubenswrapper[4808]: I1002 16:43:01.961039 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sksnr"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.003761 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.003842 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.004061 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmglv"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.009448 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.009549 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.025821 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmglv" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.029875 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmfsh"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.030338 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sksnr" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.031422 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ce081a7-e90f-40ac-aa28-5ef6e19af29a-serving-cert\") pod \"openshift-config-operator-7777fb866f-lbw7q\" (UID: \"0ce081a7-e90f-40ac-aa28-5ef6e19af29a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lbw7q" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.031640 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.031792 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.031852 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.032014 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.032077 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.032244 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.032323 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.032344 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.032537 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.033204 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctxzc\" (UniqueName: \"kubernetes.io/projected/8bc8bd65-c754-4a0a-9f11-40d444e2492a-kube-api-access-ctxzc\") pod \"downloads-7954f5f757-qsqkp\" (UID: \"8bc8bd65-c754-4a0a-9f11-40d444e2492a\") " pod="openshift-console/downloads-7954f5f757-qsqkp" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.033402 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppx9j\" (UniqueName: \"kubernetes.io/projected/44a4d77b-db5b-41ac-89b2-a2eadab424ac-kube-api-access-ppx9j\") pod \"cluster-samples-operator-665b6dd947-f5zh9\" (UID: \"44a4d77b-db5b-41ac-89b2-a2eadab424ac\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5zh9" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.033501 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0ce081a7-e90f-40ac-aa28-5ef6e19af29a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-lbw7q\" (UID: \"0ce081a7-e90f-40ac-aa28-5ef6e19af29a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lbw7q" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.033568 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/23bd8519-5b14-499f-9277-154d7397fed6-trusted-ca\") pod \"console-operator-58897d9998-t4qrz\" (UID: \"23bd8519-5b14-499f-9277-154d7397fed6\") " pod="openshift-console-operator/console-operator-58897d9998-t4qrz" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.033615 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08c9791c-01e0-4629-b121-2290fd921ff5-config\") pod \"openshift-apiserver-operator-796bbdcf4f-p9xvr\" (UID: \"08c9791c-01e0-4629-b121-2290fd921ff5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9xvr" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.033647 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08c9791c-01e0-4629-b121-2290fd921ff5-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-p9xvr\" (UID: \"08c9791c-01e0-4629-b121-2290fd921ff5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9xvr" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.033672 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/44a4d77b-db5b-41ac-89b2-a2eadab424ac-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-f5zh9\" (UID: \"44a4d77b-db5b-41ac-89b2-a2eadab424ac\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5zh9" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.033761 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23bd8519-5b14-499f-9277-154d7397fed6-config\") pod \"console-operator-58897d9998-t4qrz\" (UID: \"23bd8519-5b14-499f-9277-154d7397fed6\") " pod="openshift-console-operator/console-operator-58897d9998-t4qrz" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.033788 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23bd8519-5b14-499f-9277-154d7397fed6-serving-cert\") pod \"console-operator-58897d9998-t4qrz\" (UID: \"23bd8519-5b14-499f-9277-154d7397fed6\") " pod="openshift-console-operator/console-operator-58897d9998-t4qrz" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.033821 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vqk4\" (UniqueName: \"kubernetes.io/projected/08c9791c-01e0-4629-b121-2290fd921ff5-kube-api-access-5vqk4\") pod \"openshift-apiserver-operator-796bbdcf4f-p9xvr\" (UID: \"08c9791c-01e0-4629-b121-2290fd921ff5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9xvr" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.033892 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvhdr\" (UniqueName: \"kubernetes.io/projected/23bd8519-5b14-499f-9277-154d7397fed6-kube-api-access-pvhdr\") pod \"console-operator-58897d9998-t4qrz\" (UID: \"23bd8519-5b14-499f-9277-154d7397fed6\") " pod="openshift-console-operator/console-operator-58897d9998-t4qrz" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.033889 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.034084 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.034450 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.034467 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.034659 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.034677 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.034801 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.034854 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.034880 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.034916 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.034989 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.035061 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.035102 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.035169 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.035197 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.035214 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.035226 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.033920 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dr7j9\" (UniqueName: \"kubernetes.io/projected/0ce081a7-e90f-40ac-aa28-5ef6e19af29a-kube-api-access-dr7j9\") pod \"openshift-config-operator-7777fb866f-lbw7q\" (UID: \"0ce081a7-e90f-40ac-aa28-5ef6e19af29a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lbw7q" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.035314 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.035344 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.035410 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.035344 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.035686 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.035732 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.035737 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.035697 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.035769 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.035873 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.035885 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.035961 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.035978 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.036049 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.036098 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.036193 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.036792 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.037018 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.039521 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.039807 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.040285 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.040425 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.040562 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.041820 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2htct"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.042456 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kp2fm"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.042766 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ppxrn"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.043126 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ppxrn" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.043508 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.043694 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmfsh" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.043843 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2htct" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.043941 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-7b9zg"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.043995 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kp2fm" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.055489 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.063197 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.066144 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-tj972"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.066991 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-tj972" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.069043 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.069796 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.073282 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.073551 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.073806 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.074055 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.076927 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.077127 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.077299 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.077445 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.077586 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.077714 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.077876 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.078535 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.081544 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.081977 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.110119 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.110731 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.111049 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.111261 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.111465 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.111636 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.111774 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.111914 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.112287 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.112430 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.122047 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.145456 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.158575 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.158842 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.159205 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.160439 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9vv2b"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.161058 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.164274 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/646a1a14-e3d6-43e8-92b8-3ae7d44e96da-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sksnr\" (UID: \"646a1a14-e3d6-43e8-92b8-3ae7d44e96da\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sksnr" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.164469 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/862243ee-b010-419a-884a-0824520ae00d-serving-cert\") pod \"route-controller-manager-6576b87f9c-6br9m\" (UID: \"862243ee-b010-419a-884a-0824520ae00d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.164534 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/67d706bc-9776-4edf-98e3-7eb03ff5ef75-etcd-serving-ca\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.164556 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.164612 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctxzc\" (UniqueName: \"kubernetes.io/projected/8bc8bd65-c754-4a0a-9f11-40d444e2492a-kube-api-access-ctxzc\") pod \"downloads-7954f5f757-qsqkp\" (UID: \"8bc8bd65-c754-4a0a-9f11-40d444e2492a\") " pod="openshift-console/downloads-7954f5f757-qsqkp" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.164728 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppx9j\" (UniqueName: \"kubernetes.io/projected/44a4d77b-db5b-41ac-89b2-a2eadab424ac-kube-api-access-ppx9j\") pod \"cluster-samples-operator-665b6dd947-f5zh9\" (UID: \"44a4d77b-db5b-41ac-89b2-a2eadab424ac\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5zh9" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.164749 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s754\" (UniqueName: \"kubernetes.io/projected/1856b6cf-f125-4641-a31f-7795b811e2c0-kube-api-access-5s754\") pod \"openshift-controller-manager-operator-756b6f6bc6-4dk9g\" (UID: \"1856b6cf-f125-4641-a31f-7795b811e2c0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4dk9g" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.164768 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/67d706bc-9776-4edf-98e3-7eb03ff5ef75-audit\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.164805 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-p47w4"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.164928 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-console-config\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.165014 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.165055 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.165216 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0ce081a7-e90f-40ac-aa28-5ef6e19af29a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-lbw7q\" (UID: \"0ce081a7-e90f-40ac-aa28-5ef6e19af29a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lbw7q" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.165341 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/862243ee-b010-419a-884a-0824520ae00d-config\") pod \"route-controller-manager-6576b87f9c-6br9m\" (UID: \"862243ee-b010-419a-884a-0824520ae00d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.165375 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/862243ee-b010-419a-884a-0824520ae00d-client-ca\") pod \"route-controller-manager-6576b87f9c-6br9m\" (UID: \"862243ee-b010-419a-884a-0824520ae00d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.165409 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/67d706bc-9776-4edf-98e3-7eb03ff5ef75-audit-dir\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.165435 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.165533 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-oauth-serving-cert\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.165550 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e-images\") pod \"machine-api-operator-5694c8668f-zqnpj\" (UID: \"3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zqnpj" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.165560 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-v4bld"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.165597 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67d706bc-9776-4edf-98e3-7eb03ff5ef75-config\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.165601 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p47w4" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.165680 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/23bd8519-5b14-499f-9277-154d7397fed6-trusted-ca\") pod \"console-operator-58897d9998-t4qrz\" (UID: \"23bd8519-5b14-499f-9277-154d7397fed6\") " pod="openshift-console-operator/console-operator-58897d9998-t4qrz" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.165922 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08c9791c-01e0-4629-b121-2290fd921ff5-config\") pod \"openshift-apiserver-operator-796bbdcf4f-p9xvr\" (UID: \"08c9791c-01e0-4629-b121-2290fd921ff5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9xvr" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.165998 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0ce081a7-e90f-40ac-aa28-5ef6e19af29a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-lbw7q\" (UID: \"0ce081a7-e90f-40ac-aa28-5ef6e19af29a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lbw7q" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.166011 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv5c8\" (UniqueName: \"kubernetes.io/projected/c58d3855-bb32-4bb3-94ae-53c49209655e-kube-api-access-lv5c8\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.166062 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205-machine-approver-tls\") pod \"machine-approver-56656f9798-lkxxm\" (UID: \"2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-lkxxm" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.166128 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08c9791c-01e0-4629-b121-2290fd921ff5-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-p9xvr\" (UID: \"08c9791c-01e0-4629-b121-2290fd921ff5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9xvr" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.166167 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/44a4d77b-db5b-41ac-89b2-a2eadab424ac-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-f5zh9\" (UID: \"44a4d77b-db5b-41ac-89b2-a2eadab424ac\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5zh9" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.166377 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4bld" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.166806 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c58d3855-bb32-4bb3-94ae-53c49209655e-console-serving-cert\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.166837 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08c9791c-01e0-4629-b121-2290fd921ff5-config\") pod \"openshift-apiserver-operator-796bbdcf4f-p9xvr\" (UID: \"08c9791c-01e0-4629-b121-2290fd921ff5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9xvr" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.166891 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/646a1a14-e3d6-43e8-92b8-3ae7d44e96da-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sksnr\" (UID: \"646a1a14-e3d6-43e8-92b8-3ae7d44e96da\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sksnr" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.166938 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/67d706bc-9776-4edf-98e3-7eb03ff5ef75-encryption-config\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.166961 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdvvn\" (UniqueName: \"kubernetes.io/projected/3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e-kube-api-access-wdvvn\") pod \"machine-api-operator-5694c8668f-zqnpj\" (UID: \"3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zqnpj" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.166983 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e615c64c-94fd-44fa-a119-69435c37bd82-etcd-client\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167012 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167025 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/23bd8519-5b14-499f-9277-154d7397fed6-trusted-ca\") pod \"console-operator-58897d9998-t4qrz\" (UID: \"23bd8519-5b14-499f-9277-154d7397fed6\") " pod="openshift-console-operator/console-operator-58897d9998-t4qrz" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167060 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167083 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkxxv\" (UniqueName: \"kubernetes.io/projected/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-kube-api-access-zkxxv\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167167 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzq84\" (UniqueName: \"kubernetes.io/projected/862243ee-b010-419a-884a-0824520ae00d-kube-api-access-bzq84\") pod \"route-controller-manager-6576b87f9c-6br9m\" (UID: \"862243ee-b010-419a-884a-0824520ae00d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167172 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167191 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167209 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167212 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zqnpj\" (UID: \"3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zqnpj" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167273 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43468992-0b2c-47ae-8293-609319017a71-config\") pod \"authentication-operator-69f744f599-ffv8w\" (UID: \"43468992-0b2c-47ae-8293-609319017a71\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ffv8w" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167289 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167307 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167330 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whqgx\" (UniqueName: \"kubernetes.io/projected/e615c64c-94fd-44fa-a119-69435c37bd82-kube-api-access-whqgx\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167348 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43468992-0b2c-47ae-8293-609319017a71-serving-cert\") pod \"authentication-operator-69f744f599-ffv8w\" (UID: \"43468992-0b2c-47ae-8293-609319017a71\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ffv8w" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167374 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e-config\") pod \"machine-api-operator-5694c8668f-zqnpj\" (UID: \"3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zqnpj" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167392 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e615c64c-94fd-44fa-a119-69435c37bd82-audit-policies\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167416 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43468992-0b2c-47ae-8293-609319017a71-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ffv8w\" (UID: \"43468992-0b2c-47ae-8293-609319017a71\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ffv8w" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167450 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-service-ca\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167486 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/646a1a14-e3d6-43e8-92b8-3ae7d44e96da-config\") pod \"kube-controller-manager-operator-78b949d7b-sksnr\" (UID: \"646a1a14-e3d6-43e8-92b8-3ae7d44e96da\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sksnr" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167537 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23bd8519-5b14-499f-9277-154d7397fed6-config\") pod \"console-operator-58897d9998-t4qrz\" (UID: \"23bd8519-5b14-499f-9277-154d7397fed6\") " pod="openshift-console-operator/console-operator-58897d9998-t4qrz" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167574 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23bd8519-5b14-499f-9277-154d7397fed6-serving-cert\") pod \"console-operator-58897d9998-t4qrz\" (UID: \"23bd8519-5b14-499f-9277-154d7397fed6\") " pod="openshift-console-operator/console-operator-58897d9998-t4qrz" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167726 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167928 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vqk4\" (UniqueName: \"kubernetes.io/projected/08c9791c-01e0-4629-b121-2290fd921ff5-kube-api-access-5vqk4\") pod \"openshift-apiserver-operator-796bbdcf4f-p9xvr\" (UID: \"08c9791c-01e0-4629-b121-2290fd921ff5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9xvr" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.167997 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1856b6cf-f125-4641-a31f-7795b811e2c0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4dk9g\" (UID: \"1856b6cf-f125-4641-a31f-7795b811e2c0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4dk9g" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168024 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e615c64c-94fd-44fa-a119-69435c37bd82-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168140 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23bd8519-5b14-499f-9277-154d7397fed6-config\") pod \"console-operator-58897d9998-t4qrz\" (UID: \"23bd8519-5b14-499f-9277-154d7397fed6\") " pod="openshift-console-operator/console-operator-58897d9998-t4qrz" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168173 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvhdr\" (UniqueName: \"kubernetes.io/projected/23bd8519-5b14-499f-9277-154d7397fed6-kube-api-access-pvhdr\") pod \"console-operator-58897d9998-t4qrz\" (UID: \"23bd8519-5b14-499f-9277-154d7397fed6\") " pod="openshift-console-operator/console-operator-58897d9998-t4qrz" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168197 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dr7j9\" (UniqueName: \"kubernetes.io/projected/0ce081a7-e90f-40ac-aa28-5ef6e19af29a-kube-api-access-dr7j9\") pod \"openshift-config-operator-7777fb866f-lbw7q\" (UID: \"0ce081a7-e90f-40ac-aa28-5ef6e19af29a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lbw7q" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168217 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/67d706bc-9776-4edf-98e3-7eb03ff5ef75-node-pullsecrets\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168302 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/67d706bc-9776-4edf-98e3-7eb03ff5ef75-image-import-ca\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168323 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168347 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168369 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ce081a7-e90f-40ac-aa28-5ef6e19af29a-serving-cert\") pod \"openshift-config-operator-7777fb866f-lbw7q\" (UID: \"0ce081a7-e90f-40ac-aa28-5ef6e19af29a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lbw7q" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168393 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q54jg\" (UniqueName: \"kubernetes.io/projected/2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205-kube-api-access-q54jg\") pod \"machine-approver-56656f9798-lkxxm\" (UID: \"2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-lkxxm" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168416 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-trusted-ca-bundle\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168433 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1856b6cf-f125-4641-a31f-7795b811e2c0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4dk9g\" (UID: \"1856b6cf-f125-4641-a31f-7795b811e2c0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4dk9g" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168450 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67d706bc-9776-4edf-98e3-7eb03ff5ef75-serving-cert\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168470 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e615c64c-94fd-44fa-a119-69435c37bd82-audit-dir\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168489 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c58d3855-bb32-4bb3-94ae-53c49209655e-console-oauth-config\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168506 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e615c64c-94fd-44fa-a119-69435c37bd82-encryption-config\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168526 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205-config\") pod \"machine-approver-56656f9798-lkxxm\" (UID: \"2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-lkxxm" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168545 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fhj2\" (UniqueName: \"kubernetes.io/projected/67d706bc-9776-4edf-98e3-7eb03ff5ef75-kube-api-access-6fhj2\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168566 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e615c64c-94fd-44fa-a119-69435c37bd82-serving-cert\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168583 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43468992-0b2c-47ae-8293-609319017a71-service-ca-bundle\") pod \"authentication-operator-69f744f599-ffv8w\" (UID: \"43468992-0b2c-47ae-8293-609319017a71\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ffv8w" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168600 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqrr9\" (UniqueName: \"kubernetes.io/projected/43468992-0b2c-47ae-8293-609319017a71-kube-api-access-kqrr9\") pod \"authentication-operator-69f744f599-ffv8w\" (UID: \"43468992-0b2c-47ae-8293-609319017a71\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ffv8w" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168623 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-audit-policies\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168639 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-audit-dir\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168659 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168677 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/67d706bc-9776-4edf-98e3-7eb03ff5ef75-etcd-client\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168693 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e615c64c-94fd-44fa-a119-69435c37bd82-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168731 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205-auth-proxy-config\") pod \"machine-approver-56656f9798-lkxxm\" (UID: \"2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-lkxxm" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.168753 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67d706bc-9776-4edf-98e3-7eb03ff5ef75-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.169654 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.170870 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qc629"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.171318 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.171328 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.171924 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ffv8w"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.172929 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4nnw"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.173771 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4nnw" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.174016 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zqnpj"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.175182 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-q7ck8"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.175295 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/44a4d77b-db5b-41ac-89b2-a2eadab424ac-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-f5zh9\" (UID: \"44a4d77b-db5b-41ac-89b2-a2eadab424ac\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5zh9" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.175432 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23bd8519-5b14-499f-9277-154d7397fed6-serving-cert\") pod \"console-operator-58897d9998-t4qrz\" (UID: \"23bd8519-5b14-499f-9277-154d7397fed6\") " pod="openshift-console-operator/console-operator-58897d9998-t4qrz" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.175691 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08c9791c-01e0-4629-b121-2290fd921ff5-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-p9xvr\" (UID: \"08c9791c-01e0-4629-b121-2290fd921ff5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9xvr" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.175699 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.175790 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ddlqv"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.175896 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ce081a7-e90f-40ac-aa28-5ef6e19af29a-serving-cert\") pod \"openshift-config-operator-7777fb866f-lbw7q\" (UID: \"0ce081a7-e90f-40ac-aa28-5ef6e19af29a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lbw7q" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.177813 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r4vs4"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.178442 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.178632 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-ddlqv" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.178805 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gpxn2"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.179052 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r4vs4" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.179175 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.179192 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.179990 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.180788 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.180992 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-7pnwt"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.181384 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-7pnwt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.184605 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.186004 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.187085 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-4ff6b"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.187607 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-4ff6b" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.190207 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.190629 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.192684 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.192809 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.196323 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmglv"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.199106 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4dk9g"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.199992 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.201671 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-bxvk8"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.202610 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2htks"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.203867 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-t4qrz"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.204022 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2htks" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.204336 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bxvk8" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.204921 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-dlddx"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.205830 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kp2fm"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.206882 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-qsqkp"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.208440 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9xvr"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.208651 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-p47w4"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.209708 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sksnr"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.212136 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.214526 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-tj972"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.216314 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2bgkb"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.216514 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8974h"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.218317 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4nnw"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.227083 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9vv2b"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.235939 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-lbw7q"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.236177 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.240484 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5zh9"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.243563 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.244888 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"04ef26a93857c9e7093be590fa17b4059fc45ca460280432ff12312b21adbed3"} Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.248171 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f2010fd861a785836d24aae48bffcc50cc08c1cc4a0b811e434f55ae5e13ef27"} Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.248263 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.248949 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ppxrn"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.252285 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.252586 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.254870 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-4ff6b"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.256605 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.259582 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-v4bld"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.261083 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2htct"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.264638 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmfsh"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.266094 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gpxn2"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.267161 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ddlqv"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.268354 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-q7ck8"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.269506 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r4vs4"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.269625 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1856b6cf-f125-4641-a31f-7795b811e2c0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4dk9g\" (UID: \"1856b6cf-f125-4641-a31f-7795b811e2c0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4dk9g" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.269683 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e615c64c-94fd-44fa-a119-69435c37bd82-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.269712 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmqj6\" (UniqueName: \"kubernetes.io/projected/9ebb9933-4652-4152-a529-5489fd9d5e5d-kube-api-access-kmqj6\") pod \"router-default-5444994796-7b9zg\" (UID: \"9ebb9933-4652-4152-a529-5489fd9d5e5d\") " pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.269733 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13d94271-250f-49f2-b1ef-4a97e991a124-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kp2fm\" (UID: \"13d94271-250f-49f2-b1ef-4a97e991a124\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kp2fm" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.269755 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhjq9\" (UniqueName: \"kubernetes.io/projected/f6a13086-47b7-4624-bd50-773a4badcb3f-kube-api-access-lhjq9\") pod \"migrator-59844c95c7-6vlpf\" (UID: \"f6a13086-47b7-4624-bd50-773a4badcb3f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.269776 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9ebb9933-4652-4152-a529-5489fd9d5e5d-default-certificate\") pod \"router-default-5444994796-7b9zg\" (UID: \"9ebb9933-4652-4152-a529-5489fd9d5e5d\") " pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.269809 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/67d706bc-9776-4edf-98e3-7eb03ff5ef75-node-pullsecrets\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.269830 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/67d706bc-9776-4edf-98e3-7eb03ff5ef75-image-import-ca\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.269847 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.269868 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bf03d573-aaae-4b60-b285-da3bb7d3e4e3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zmglv\" (UID: \"bf03d573-aaae-4b60-b285-da3bb7d3e4e3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmglv" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.269890 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.269913 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ae1cc03-9560-4e85-889e-63eda4518302-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-2htct\" (UID: \"6ae1cc03-9560-4e85-889e-63eda4518302\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2htct" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.269934 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/580bbc4a-b873-4631-86c3-476a771b9a70-profile-collector-cert\") pod \"catalog-operator-68c6474976-ppxrn\" (UID: \"580bbc4a-b873-4631-86c3-476a771b9a70\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ppxrn" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.269955 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q54jg\" (UniqueName: \"kubernetes.io/projected/2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205-kube-api-access-q54jg\") pod \"machine-approver-56656f9798-lkxxm\" (UID: \"2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-lkxxm" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270006 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-trusted-ca-bundle\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270024 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1856b6cf-f125-4641-a31f-7795b811e2c0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4dk9g\" (UID: \"1856b6cf-f125-4641-a31f-7795b811e2c0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4dk9g" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270051 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67d706bc-9776-4edf-98e3-7eb03ff5ef75-serving-cert\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270072 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/67d706bc-9776-4edf-98e3-7eb03ff5ef75-node-pullsecrets\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270082 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9ebb9933-4652-4152-a529-5489fd9d5e5d-stats-auth\") pod \"router-default-5444994796-7b9zg\" (UID: \"9ebb9933-4652-4152-a529-5489fd9d5e5d\") " pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270206 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/05b1a2b9-4809-4e5b-87e3-49b61d3c5184-trusted-ca\") pod \"ingress-operator-5b745b69d9-zxfsc\" (UID: \"05b1a2b9-4809-4e5b-87e3-49b61d3c5184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270265 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c58d3855-bb32-4bb3-94ae-53c49209655e-console-oauth-config\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270325 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e615c64c-94fd-44fa-a119-69435c37bd82-encryption-config\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270360 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e615c64c-94fd-44fa-a119-69435c37bd82-audit-dir\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270411 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ebb9933-4652-4152-a529-5489fd9d5e5d-metrics-certs\") pod \"router-default-5444994796-7b9zg\" (UID: \"9ebb9933-4652-4152-a529-5489fd9d5e5d\") " pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270447 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205-config\") pod \"machine-approver-56656f9798-lkxxm\" (UID: \"2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-lkxxm" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270509 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fhj2\" (UniqueName: \"kubernetes.io/projected/67d706bc-9776-4edf-98e3-7eb03ff5ef75-kube-api-access-6fhj2\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270543 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e615c64c-94fd-44fa-a119-69435c37bd82-serving-cert\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270573 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43468992-0b2c-47ae-8293-609319017a71-service-ca-bundle\") pod \"authentication-operator-69f744f599-ffv8w\" (UID: \"43468992-0b2c-47ae-8293-609319017a71\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ffv8w" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270598 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqrr9\" (UniqueName: \"kubernetes.io/projected/43468992-0b2c-47ae-8293-609319017a71-kube-api-access-kqrr9\") pod \"authentication-operator-69f744f599-ffv8w\" (UID: \"43468992-0b2c-47ae-8293-609319017a71\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ffv8w" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270618 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-audit-policies\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270643 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13d94271-250f-49f2-b1ef-4a97e991a124-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kp2fm\" (UID: \"13d94271-250f-49f2-b1ef-4a97e991a124\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kp2fm" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270666 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/67d706bc-9776-4edf-98e3-7eb03ff5ef75-etcd-client\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270685 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e615c64c-94fd-44fa-a119-69435c37bd82-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270706 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-audit-dir\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270745 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270774 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205-auth-proxy-config\") pod \"machine-approver-56656f9798-lkxxm\" (UID: \"2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-lkxxm" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270799 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67d706bc-9776-4edf-98e3-7eb03ff5ef75-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270819 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/580bbc4a-b873-4631-86c3-476a771b9a70-srv-cert\") pod \"catalog-operator-68c6474976-ppxrn\" (UID: \"580bbc4a-b873-4631-86c3-476a771b9a70\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ppxrn" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270840 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s78xt\" (UniqueName: \"kubernetes.io/projected/580bbc4a-b873-4631-86c3-476a771b9a70-kube-api-access-s78xt\") pod \"catalog-operator-68c6474976-ppxrn\" (UID: \"580bbc4a-b873-4631-86c3-476a771b9a70\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ppxrn" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270887 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/646a1a14-e3d6-43e8-92b8-3ae7d44e96da-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sksnr\" (UID: \"646a1a14-e3d6-43e8-92b8-3ae7d44e96da\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sksnr" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270910 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/862243ee-b010-419a-884a-0824520ae00d-serving-cert\") pod \"route-controller-manager-6576b87f9c-6br9m\" (UID: \"862243ee-b010-419a-884a-0824520ae00d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270929 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/67d706bc-9776-4edf-98e3-7eb03ff5ef75-etcd-serving-ca\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270951 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270973 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-console-config\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.270991 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s754\" (UniqueName: \"kubernetes.io/projected/1856b6cf-f125-4641-a31f-7795b811e2c0-kube-api-access-5s754\") pod \"openshift-controller-manager-operator-756b6f6bc6-4dk9g\" (UID: \"1856b6cf-f125-4641-a31f-7795b811e2c0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4dk9g" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271010 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/67d706bc-9776-4edf-98e3-7eb03ff5ef75-audit\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271030 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/13d94271-250f-49f2-b1ef-4a97e991a124-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kp2fm\" (UID: \"13d94271-250f-49f2-b1ef-4a97e991a124\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kp2fm" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271053 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271078 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/862243ee-b010-419a-884a-0824520ae00d-config\") pod \"route-controller-manager-6576b87f9c-6br9m\" (UID: \"862243ee-b010-419a-884a-0824520ae00d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271094 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/862243ee-b010-419a-884a-0824520ae00d-client-ca\") pod \"route-controller-manager-6576b87f9c-6br9m\" (UID: \"862243ee-b010-419a-884a-0824520ae00d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271113 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/05b1a2b9-4809-4e5b-87e3-49b61d3c5184-metrics-tls\") pod \"ingress-operator-5b745b69d9-zxfsc\" (UID: \"05b1a2b9-4809-4e5b-87e3-49b61d3c5184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271141 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/67d706bc-9776-4edf-98e3-7eb03ff5ef75-audit-dir\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271140 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271159 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271301 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7777a7df-de5d-4287-a6e9-bcb70a07fdfd-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-tj972\" (UID: \"7777a7df-de5d-4287-a6e9-bcb70a07fdfd\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tj972" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271330 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/05b1a2b9-4809-4e5b-87e3-49b61d3c5184-bound-sa-token\") pod \"ingress-operator-5b745b69d9-zxfsc\" (UID: \"05b1a2b9-4809-4e5b-87e3-49b61d3c5184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271356 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnxsg\" (UniqueName: \"kubernetes.io/projected/05b1a2b9-4809-4e5b-87e3-49b61d3c5184-kube-api-access-qnxsg\") pod \"ingress-operator-5b745b69d9-zxfsc\" (UID: \"05b1a2b9-4809-4e5b-87e3-49b61d3c5184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271360 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/67d706bc-9776-4edf-98e3-7eb03ff5ef75-image-import-ca\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271385 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-oauth-serving-cert\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271420 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e-images\") pod \"machine-api-operator-5694c8668f-zqnpj\" (UID: \"3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zqnpj" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271438 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67d706bc-9776-4edf-98e3-7eb03ff5ef75-config\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271456 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf03d573-aaae-4b60-b285-da3bb7d3e4e3-config\") pod \"kube-apiserver-operator-766d6c64bb-zmglv\" (UID: \"bf03d573-aaae-4b60-b285-da3bb7d3e4e3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmglv" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271479 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv5c8\" (UniqueName: \"kubernetes.io/projected/c58d3855-bb32-4bb3-94ae-53c49209655e-kube-api-access-lv5c8\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271503 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c58d3855-bb32-4bb3-94ae-53c49209655e-console-serving-cert\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271524 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/646a1a14-e3d6-43e8-92b8-3ae7d44e96da-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sksnr\" (UID: \"646a1a14-e3d6-43e8-92b8-3ae7d44e96da\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sksnr" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271547 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205-machine-approver-tls\") pod \"machine-approver-56656f9798-lkxxm\" (UID: \"2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-lkxxm" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271580 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/67d706bc-9776-4edf-98e3-7eb03ff5ef75-encryption-config\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271608 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdvvn\" (UniqueName: \"kubernetes.io/projected/3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e-kube-api-access-wdvvn\") pod \"machine-api-operator-5694c8668f-zqnpj\" (UID: \"3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zqnpj" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271628 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e615c64c-94fd-44fa-a119-69435c37bd82-etcd-client\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271645 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271672 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271691 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkxxv\" (UniqueName: \"kubernetes.io/projected/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-kube-api-access-zkxxv\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271724 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzq84\" (UniqueName: \"kubernetes.io/projected/862243ee-b010-419a-884a-0824520ae00d-kube-api-access-bzq84\") pod \"route-controller-manager-6576b87f9c-6br9m\" (UID: \"862243ee-b010-419a-884a-0824520ae00d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271743 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zqnpj\" (UID: \"3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zqnpj" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271760 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43468992-0b2c-47ae-8293-609319017a71-config\") pod \"authentication-operator-69f744f599-ffv8w\" (UID: \"43468992-0b2c-47ae-8293-609319017a71\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ffv8w" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271779 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271786 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43468992-0b2c-47ae-8293-609319017a71-service-ca-bundle\") pod \"authentication-operator-69f744f599-ffv8w\" (UID: \"43468992-0b2c-47ae-8293-609319017a71\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ffv8w" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271796 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271824 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271848 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whqgx\" (UniqueName: \"kubernetes.io/projected/e615c64c-94fd-44fa-a119-69435c37bd82-kube-api-access-whqgx\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271869 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43468992-0b2c-47ae-8293-609319017a71-serving-cert\") pod \"authentication-operator-69f744f599-ffv8w\" (UID: \"43468992-0b2c-47ae-8293-609319017a71\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ffv8w" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271893 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb48v\" (UniqueName: \"kubernetes.io/projected/6ae1cc03-9560-4e85-889e-63eda4518302-kube-api-access-xb48v\") pod \"kube-storage-version-migrator-operator-b67b599dd-2htct\" (UID: \"6ae1cc03-9560-4e85-889e-63eda4518302\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2htct" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271923 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e-config\") pod \"machine-api-operator-5694c8668f-zqnpj\" (UID: \"3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zqnpj" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271943 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e615c64c-94fd-44fa-a119-69435c37bd82-audit-policies\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271960 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43468992-0b2c-47ae-8293-609319017a71-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ffv8w\" (UID: \"43468992-0b2c-47ae-8293-609319017a71\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ffv8w" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.271986 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-service-ca\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.272008 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/646a1a14-e3d6-43e8-92b8-3ae7d44e96da-config\") pod \"kube-controller-manager-operator-78b949d7b-sksnr\" (UID: \"646a1a14-e3d6-43e8-92b8-3ae7d44e96da\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sksnr" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.272025 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ae1cc03-9560-4e85-889e-63eda4518302-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-2htct\" (UID: \"6ae1cc03-9560-4e85-889e-63eda4518302\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2htct" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.272062 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9qhp\" (UniqueName: \"kubernetes.io/projected/7777a7df-de5d-4287-a6e9-bcb70a07fdfd-kube-api-access-z9qhp\") pod \"multus-admission-controller-857f4d67dd-tj972\" (UID: \"7777a7df-de5d-4287-a6e9-bcb70a07fdfd\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tj972" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.272078 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9ebb9933-4652-4152-a529-5489fd9d5e5d-service-ca-bundle\") pod \"router-default-5444994796-7b9zg\" (UID: \"9ebb9933-4652-4152-a529-5489fd9d5e5d\") " pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.272096 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf03d573-aaae-4b60-b285-da3bb7d3e4e3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zmglv\" (UID: \"bf03d573-aaae-4b60-b285-da3bb7d3e4e3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmglv" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.272402 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.273062 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-audit-policies\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.273072 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.273310 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205-config\") pod \"machine-approver-56656f9798-lkxxm\" (UID: \"2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-lkxxm" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.274474 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-oauth-serving-cert\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.274698 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e615c64c-94fd-44fa-a119-69435c37bd82-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.274825 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e615c64c-94fd-44fa-a119-69435c37bd82-audit-dir\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.274959 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1856b6cf-f125-4641-a31f-7795b811e2c0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4dk9g\" (UID: \"1856b6cf-f125-4641-a31f-7795b811e2c0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4dk9g" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.275378 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e-images\") pod \"machine-api-operator-5694c8668f-zqnpj\" (UID: \"3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zqnpj" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.275526 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1856b6cf-f125-4641-a31f-7795b811e2c0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4dk9g\" (UID: \"1856b6cf-f125-4641-a31f-7795b811e2c0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4dk9g" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.275528 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.276275 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c58d3855-bb32-4bb3-94ae-53c49209655e-console-oauth-config\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.276584 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-trusted-ca-bundle\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.276642 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qc629"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.277129 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e615c64c-94fd-44fa-a119-69435c37bd82-encryption-config\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.277533 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/67d706bc-9776-4edf-98e3-7eb03ff5ef75-etcd-client\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.277544 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-console-config\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.277624 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/67d706bc-9776-4edf-98e3-7eb03ff5ef75-audit-dir\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.277803 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/67d706bc-9776-4edf-98e3-7eb03ff5ef75-audit\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.278086 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/862243ee-b010-419a-884a-0824520ae00d-client-ca\") pod \"route-controller-manager-6576b87f9c-6br9m\" (UID: \"862243ee-b010-419a-884a-0824520ae00d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.278399 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e615c64c-94fd-44fa-a119-69435c37bd82-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.278625 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-audit-dir\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.278869 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/67d706bc-9776-4edf-98e3-7eb03ff5ef75-etcd-serving-ca\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.278879 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67d706bc-9776-4edf-98e3-7eb03ff5ef75-serving-cert\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.278968 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43468992-0b2c-47ae-8293-609319017a71-config\") pod \"authentication-operator-69f744f599-ffv8w\" (UID: \"43468992-0b2c-47ae-8293-609319017a71\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ffv8w" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.278994 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-vh4gp"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.280419 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-vh4gp"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.280647 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-vh4gp" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.281546 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/646a1a14-e3d6-43e8-92b8-3ae7d44e96da-config\") pod \"kube-controller-manager-operator-78b949d7b-sksnr\" (UID: \"646a1a14-e3d6-43e8-92b8-3ae7d44e96da\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sksnr" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.281572 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205-auth-proxy-config\") pod \"machine-approver-56656f9798-lkxxm\" (UID: \"2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-lkxxm" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.281101 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e615c64c-94fd-44fa-a119-69435c37bd82-audit-policies\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.280835 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/862243ee-b010-419a-884a-0824520ae00d-config\") pod \"route-controller-manager-6576b87f9c-6br9m\" (UID: \"862243ee-b010-419a-884a-0824520ae00d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.282419 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.282444 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67d706bc-9776-4edf-98e3-7eb03ff5ef75-config\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.282512 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.282737 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e-config\") pod \"machine-api-operator-5694c8668f-zqnpj\" (UID: \"3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zqnpj" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.282864 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67d706bc-9776-4edf-98e3-7eb03ff5ef75-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.283039 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.283465 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.283870 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43468992-0b2c-47ae-8293-609319017a71-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ffv8w\" (UID: \"43468992-0b2c-47ae-8293-609319017a71\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ffv8w" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.284280 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-service-ca\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.284595 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.285338 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/862243ee-b010-419a-884a-0824520ae00d-serving-cert\") pod \"route-controller-manager-6576b87f9c-6br9m\" (UID: \"862243ee-b010-419a-884a-0824520ae00d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.285870 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.285870 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.286393 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/646a1a14-e3d6-43e8-92b8-3ae7d44e96da-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sksnr\" (UID: \"646a1a14-e3d6-43e8-92b8-3ae7d44e96da\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sksnr" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.286509 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/67d706bc-9776-4edf-98e3-7eb03ff5ef75-encryption-config\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.287104 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205-machine-approver-tls\") pod \"machine-approver-56656f9798-lkxxm\" (UID: \"2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-lkxxm" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.289342 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-7pnwt"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.289645 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43468992-0b2c-47ae-8293-609319017a71-serving-cert\") pod \"authentication-operator-69f744f599-ffv8w\" (UID: \"43468992-0b2c-47ae-8293-609319017a71\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ffv8w" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.289714 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.290786 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e615c64c-94fd-44fa-a119-69435c37bd82-etcd-client\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.291749 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c58d3855-bb32-4bb3-94ae-53c49209655e-console-serving-cert\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.291870 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2htks"] Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.293510 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.293759 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.294091 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.294160 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.299660 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e615c64c-94fd-44fa-a119-69435c37bd82-serving-cert\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.304128 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zqnpj\" (UID: \"3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zqnpj" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.314217 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.333877 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.352921 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.372774 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.372961 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9ebb9933-4652-4152-a529-5489fd9d5e5d-default-certificate\") pod \"router-default-5444994796-7b9zg\" (UID: \"9ebb9933-4652-4152-a529-5489fd9d5e5d\") " pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373029 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bf03d573-aaae-4b60-b285-da3bb7d3e4e3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zmglv\" (UID: \"bf03d573-aaae-4b60-b285-da3bb7d3e4e3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmglv" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373061 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ae1cc03-9560-4e85-889e-63eda4518302-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-2htct\" (UID: \"6ae1cc03-9560-4e85-889e-63eda4518302\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2htct" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373092 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/580bbc4a-b873-4631-86c3-476a771b9a70-profile-collector-cert\") pod \"catalog-operator-68c6474976-ppxrn\" (UID: \"580bbc4a-b873-4631-86c3-476a771b9a70\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ppxrn" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373130 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/05b1a2b9-4809-4e5b-87e3-49b61d3c5184-trusted-ca\") pod \"ingress-operator-5b745b69d9-zxfsc\" (UID: \"05b1a2b9-4809-4e5b-87e3-49b61d3c5184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373159 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9ebb9933-4652-4152-a529-5489fd9d5e5d-stats-auth\") pod \"router-default-5444994796-7b9zg\" (UID: \"9ebb9933-4652-4152-a529-5489fd9d5e5d\") " pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373182 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ebb9933-4652-4152-a529-5489fd9d5e5d-metrics-certs\") pod \"router-default-5444994796-7b9zg\" (UID: \"9ebb9933-4652-4152-a529-5489fd9d5e5d\") " pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373258 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13d94271-250f-49f2-b1ef-4a97e991a124-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kp2fm\" (UID: \"13d94271-250f-49f2-b1ef-4a97e991a124\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kp2fm" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373287 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/580bbc4a-b873-4631-86c3-476a771b9a70-srv-cert\") pod \"catalog-operator-68c6474976-ppxrn\" (UID: \"580bbc4a-b873-4631-86c3-476a771b9a70\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ppxrn" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373312 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s78xt\" (UniqueName: \"kubernetes.io/projected/580bbc4a-b873-4631-86c3-476a771b9a70-kube-api-access-s78xt\") pod \"catalog-operator-68c6474976-ppxrn\" (UID: \"580bbc4a-b873-4631-86c3-476a771b9a70\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ppxrn" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373374 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/13d94271-250f-49f2-b1ef-4a97e991a124-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kp2fm\" (UID: \"13d94271-250f-49f2-b1ef-4a97e991a124\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kp2fm" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373418 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/05b1a2b9-4809-4e5b-87e3-49b61d3c5184-metrics-tls\") pod \"ingress-operator-5b745b69d9-zxfsc\" (UID: \"05b1a2b9-4809-4e5b-87e3-49b61d3c5184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373443 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnxsg\" (UniqueName: \"kubernetes.io/projected/05b1a2b9-4809-4e5b-87e3-49b61d3c5184-kube-api-access-qnxsg\") pod \"ingress-operator-5b745b69d9-zxfsc\" (UID: \"05b1a2b9-4809-4e5b-87e3-49b61d3c5184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373478 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7777a7df-de5d-4287-a6e9-bcb70a07fdfd-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-tj972\" (UID: \"7777a7df-de5d-4287-a6e9-bcb70a07fdfd\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tj972" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373501 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/05b1a2b9-4809-4e5b-87e3-49b61d3c5184-bound-sa-token\") pod \"ingress-operator-5b745b69d9-zxfsc\" (UID: \"05b1a2b9-4809-4e5b-87e3-49b61d3c5184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373528 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf03d573-aaae-4b60-b285-da3bb7d3e4e3-config\") pod \"kube-apiserver-operator-766d6c64bb-zmglv\" (UID: \"bf03d573-aaae-4b60-b285-da3bb7d3e4e3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmglv" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373609 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb48v\" (UniqueName: \"kubernetes.io/projected/6ae1cc03-9560-4e85-889e-63eda4518302-kube-api-access-xb48v\") pod \"kube-storage-version-migrator-operator-b67b599dd-2htct\" (UID: \"6ae1cc03-9560-4e85-889e-63eda4518302\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2htct" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373638 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ae1cc03-9560-4e85-889e-63eda4518302-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-2htct\" (UID: \"6ae1cc03-9560-4e85-889e-63eda4518302\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2htct" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373665 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9qhp\" (UniqueName: \"kubernetes.io/projected/7777a7df-de5d-4287-a6e9-bcb70a07fdfd-kube-api-access-z9qhp\") pod \"multus-admission-controller-857f4d67dd-tj972\" (UID: \"7777a7df-de5d-4287-a6e9-bcb70a07fdfd\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tj972" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373690 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9ebb9933-4652-4152-a529-5489fd9d5e5d-service-ca-bundle\") pod \"router-default-5444994796-7b9zg\" (UID: \"9ebb9933-4652-4152-a529-5489fd9d5e5d\") " pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373713 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf03d573-aaae-4b60-b285-da3bb7d3e4e3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zmglv\" (UID: \"bf03d573-aaae-4b60-b285-da3bb7d3e4e3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmglv" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373736 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13d94271-250f-49f2-b1ef-4a97e991a124-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kp2fm\" (UID: \"13d94271-250f-49f2-b1ef-4a97e991a124\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kp2fm" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373769 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmqj6\" (UniqueName: \"kubernetes.io/projected/9ebb9933-4652-4152-a529-5489fd9d5e5d-kube-api-access-kmqj6\") pod \"router-default-5444994796-7b9zg\" (UID: \"9ebb9933-4652-4152-a529-5489fd9d5e5d\") " pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.373793 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhjq9\" (UniqueName: \"kubernetes.io/projected/f6a13086-47b7-4624-bd50-773a4badcb3f-kube-api-access-lhjq9\") pod \"migrator-59844c95c7-6vlpf\" (UID: \"f6a13086-47b7-4624-bd50-773a4badcb3f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.375449 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf03d573-aaae-4b60-b285-da3bb7d3e4e3-config\") pod \"kube-apiserver-operator-766d6c64bb-zmglv\" (UID: \"bf03d573-aaae-4b60-b285-da3bb7d3e4e3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmglv" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.376473 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ae1cc03-9560-4e85-889e-63eda4518302-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-2htct\" (UID: \"6ae1cc03-9560-4e85-889e-63eda4518302\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2htct" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.376949 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13d94271-250f-49f2-b1ef-4a97e991a124-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kp2fm\" (UID: \"13d94271-250f-49f2-b1ef-4a97e991a124\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kp2fm" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.377823 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ae1cc03-9560-4e85-889e-63eda4518302-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-2htct\" (UID: \"6ae1cc03-9560-4e85-889e-63eda4518302\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2htct" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.378334 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/580bbc4a-b873-4631-86c3-476a771b9a70-srv-cert\") pod \"catalog-operator-68c6474976-ppxrn\" (UID: \"580bbc4a-b873-4631-86c3-476a771b9a70\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ppxrn" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.378814 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/580bbc4a-b873-4631-86c3-476a771b9a70-profile-collector-cert\") pod \"catalog-operator-68c6474976-ppxrn\" (UID: \"580bbc4a-b873-4631-86c3-476a771b9a70\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ppxrn" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.379795 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13d94271-250f-49f2-b1ef-4a97e991a124-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kp2fm\" (UID: \"13d94271-250f-49f2-b1ef-4a97e991a124\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kp2fm" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.380012 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf03d573-aaae-4b60-b285-da3bb7d3e4e3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zmglv\" (UID: \"bf03d573-aaae-4b60-b285-da3bb7d3e4e3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmglv" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.393366 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.412893 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.416733 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9ebb9933-4652-4152-a529-5489fd9d5e5d-default-certificate\") pod \"router-default-5444994796-7b9zg\" (UID: \"9ebb9933-4652-4152-a529-5489fd9d5e5d\") " pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.432822 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.437221 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9ebb9933-4652-4152-a529-5489fd9d5e5d-stats-auth\") pod \"router-default-5444994796-7b9zg\" (UID: \"9ebb9933-4652-4152-a529-5489fd9d5e5d\") " pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.452624 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.459568 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ebb9933-4652-4152-a529-5489fd9d5e5d-metrics-certs\") pod \"router-default-5444994796-7b9zg\" (UID: \"9ebb9933-4652-4152-a529-5489fd9d5e5d\") " pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.472300 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.493204 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.513518 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.516395 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9ebb9933-4652-4152-a529-5489fd9d5e5d-service-ca-bundle\") pod \"router-default-5444994796-7b9zg\" (UID: \"9ebb9933-4652-4152-a529-5489fd9d5e5d\") " pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.532708 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.538841 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7777a7df-de5d-4287-a6e9-bcb70a07fdfd-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-tj972\" (UID: \"7777a7df-de5d-4287-a6e9-bcb70a07fdfd\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tj972" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.553705 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.573415 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.592539 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.613764 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.618825 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/05b1a2b9-4809-4e5b-87e3-49b61d3c5184-metrics-tls\") pod \"ingress-operator-5b745b69d9-zxfsc\" (UID: \"05b1a2b9-4809-4e5b-87e3-49b61d3c5184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.651698 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.654132 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.655377 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/05b1a2b9-4809-4e5b-87e3-49b61d3c5184-trusted-ca\") pod \"ingress-operator-5b745b69d9-zxfsc\" (UID: \"05b1a2b9-4809-4e5b-87e3-49b61d3c5184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.700417 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.714346 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.733468 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.753011 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.790050 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppx9j\" (UniqueName: \"kubernetes.io/projected/44a4d77b-db5b-41ac-89b2-a2eadab424ac-kube-api-access-ppx9j\") pod \"cluster-samples-operator-665b6dd947-f5zh9\" (UID: \"44a4d77b-db5b-41ac-89b2-a2eadab424ac\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5zh9" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.822380 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctxzc\" (UniqueName: \"kubernetes.io/projected/8bc8bd65-c754-4a0a-9f11-40d444e2492a-kube-api-access-ctxzc\") pod \"downloads-7954f5f757-qsqkp\" (UID: \"8bc8bd65-c754-4a0a-9f11-40d444e2492a\") " pod="openshift-console/downloads-7954f5f757-qsqkp" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.833988 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.852927 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.854004 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-qsqkp" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.873200 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.893961 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.915961 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.933738 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.953098 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.957383 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5zh9" Oct 02 16:43:02 crc kubenswrapper[4808]: I1002 16:43:02.989245 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vqk4\" (UniqueName: \"kubernetes.io/projected/08c9791c-01e0-4629-b121-2290fd921ff5-kube-api-access-5vqk4\") pod \"openshift-apiserver-operator-796bbdcf4f-p9xvr\" (UID: \"08c9791c-01e0-4629-b121-2290fd921ff5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9xvr" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.008217 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvhdr\" (UniqueName: \"kubernetes.io/projected/23bd8519-5b14-499f-9277-154d7397fed6-kube-api-access-pvhdr\") pod \"console-operator-58897d9998-t4qrz\" (UID: \"23bd8519-5b14-499f-9277-154d7397fed6\") " pod="openshift-console-operator/console-operator-58897d9998-t4qrz" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.029680 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dr7j9\" (UniqueName: \"kubernetes.io/projected/0ce081a7-e90f-40ac-aa28-5ef6e19af29a-kube-api-access-dr7j9\") pod \"openshift-config-operator-7777fb866f-lbw7q\" (UID: \"0ce081a7-e90f-40ac-aa28-5ef6e19af29a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lbw7q" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.034599 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.057672 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.072698 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-qsqkp"] Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.080110 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 02 16:43:03 crc kubenswrapper[4808]: W1002 16:43:03.081686 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8bc8bd65_c754_4a0a_9f11_40d444e2492a.slice/crio-e14527e88269ac95f55f27d84aa53c9e04e725a40d41b12cb0ce56155371f806 WatchSource:0}: Error finding container e14527e88269ac95f55f27d84aa53c9e04e725a40d41b12cb0ce56155371f806: Status 404 returned error can't find the container with id e14527e88269ac95f55f27d84aa53c9e04e725a40d41b12cb0ce56155371f806 Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.101528 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.109754 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9xvr" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.112705 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.119471 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-t4qrz" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.132302 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.143194 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5zh9"] Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.152438 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.161273 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lbw7q" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.174174 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.192063 4808 request.go:700] Waited for 1.020135314s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmco-proxy-tls&limit=500&resourceVersion=0 Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.194681 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.215047 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.232433 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.253657 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.256178 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-qsqkp" event={"ID":"8bc8bd65-c754-4a0a-9f11-40d444e2492a","Type":"ContainerStarted","Data":"e14527e88269ac95f55f27d84aa53c9e04e725a40d41b12cb0ce56155371f806"} Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.273748 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.294079 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.317107 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.324569 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-t4qrz"] Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.332746 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.354383 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.373377 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.374274 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9xvr"] Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.392590 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 02 16:43:03 crc kubenswrapper[4808]: W1002 16:43:03.401531 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08c9791c_01e0_4629_b121_2290fd921ff5.slice/crio-f6f87bbb381b038767b7986f0e281fa48bbfe499e1b27ddc55982632a0ff81f5 WatchSource:0}: Error finding container f6f87bbb381b038767b7986f0e281fa48bbfe499e1b27ddc55982632a0ff81f5: Status 404 returned error can't find the container with id f6f87bbb381b038767b7986f0e281fa48bbfe499e1b27ddc55982632a0ff81f5 Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.410435 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-lbw7q"] Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.412837 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.433391 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 02 16:43:03 crc kubenswrapper[4808]: W1002 16:43:03.440555 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ce081a7_e90f_40ac_aa28_5ef6e19af29a.slice/crio-ed1ed0cab80c5806ed19c6a8eb0dd583154d244ed8733da4f35cd1fe66273f8a WatchSource:0}: Error finding container ed1ed0cab80c5806ed19c6a8eb0dd583154d244ed8733da4f35cd1fe66273f8a: Status 404 returned error can't find the container with id ed1ed0cab80c5806ed19c6a8eb0dd583154d244ed8733da4f35cd1fe66273f8a Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.452899 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.472840 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.492827 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.513185 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.533371 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.552845 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.572878 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.593299 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.628561 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.633698 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.653086 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.673272 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.692943 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.713158 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.732782 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.753997 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.772196 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.793608 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.813739 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.833037 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.853261 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.872208 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.892782 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.913275 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.932837 4808 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.953215 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.973771 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 02 16:43:03 crc kubenswrapper[4808]: I1002 16:43:03.992337 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.032604 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqrr9\" (UniqueName: \"kubernetes.io/projected/43468992-0b2c-47ae-8293-609319017a71-kube-api-access-kqrr9\") pod \"authentication-operator-69f744f599-ffv8w\" (UID: \"43468992-0b2c-47ae-8293-609319017a71\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ffv8w" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.058280 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q54jg\" (UniqueName: \"kubernetes.io/projected/2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205-kube-api-access-q54jg\") pod \"machine-approver-56656f9798-lkxxm\" (UID: \"2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-lkxxm" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.082298 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fhj2\" (UniqueName: \"kubernetes.io/projected/67d706bc-9776-4edf-98e3-7eb03ff5ef75-kube-api-access-6fhj2\") pod \"apiserver-76f77b778f-2bgkb\" (UID: \"67d706bc-9776-4edf-98e3-7eb03ff5ef75\") " pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.098602 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s754\" (UniqueName: \"kubernetes.io/projected/1856b6cf-f125-4641-a31f-7795b811e2c0-kube-api-access-5s754\") pod \"openshift-controller-manager-operator-756b6f6bc6-4dk9g\" (UID: \"1856b6cf-f125-4641-a31f-7795b811e2c0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4dk9g" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.105552 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ffv8w" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.121825 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whqgx\" (UniqueName: \"kubernetes.io/projected/e615c64c-94fd-44fa-a119-69435c37bd82-kube-api-access-whqgx\") pod \"apiserver-7bbb656c7d-5wvmf\" (UID: \"e615c64c-94fd-44fa-a119-69435c37bd82\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.138334 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-lkxxm" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.138434 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.138568 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/646a1a14-e3d6-43e8-92b8-3ae7d44e96da-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sksnr\" (UID: \"646a1a14-e3d6-43e8-92b8-3ae7d44e96da\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sksnr" Oct 02 16:43:04 crc kubenswrapper[4808]: W1002 16:43:04.157053 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ea7e7af_a0cd_4fda_8caf_b9db0c8e5205.slice/crio-d80321049fec7b0127003afc8be505c54b9559a05fd7c37a4048da72448ce3de WatchSource:0}: Error finding container d80321049fec7b0127003afc8be505c54b9559a05fd7c37a4048da72448ce3de: Status 404 returned error can't find the container with id d80321049fec7b0127003afc8be505c54b9559a05fd7c37a4048da72448ce3de Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.166326 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4dk9g" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.168832 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv5c8\" (UniqueName: \"kubernetes.io/projected/c58d3855-bb32-4bb3-94ae-53c49209655e-kube-api-access-lv5c8\") pod \"console-f9d7485db-dlddx\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.172707 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.229477 4808 request.go:700] Waited for 1.947057147s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/serviceaccounts/oauth-openshift/token Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.236752 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.258026 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkxxv\" (UniqueName: \"kubernetes.io/projected/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-kube-api-access-zkxxv\") pod \"oauth-openshift-558db77b4-8974h\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.258874 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.260185 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzq84\" (UniqueName: \"kubernetes.io/projected/862243ee-b010-419a-884a-0824520ae00d-kube-api-access-bzq84\") pod \"route-controller-manager-6576b87f9c-6br9m\" (UID: \"862243ee-b010-419a-884a-0824520ae00d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.269385 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.274030 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdvvn\" (UniqueName: \"kubernetes.io/projected/3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e-kube-api-access-wdvvn\") pod \"machine-api-operator-5694c8668f-zqnpj\" (UID: \"3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zqnpj" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.276514 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-qsqkp" event={"ID":"8bc8bd65-c754-4a0a-9f11-40d444e2492a","Type":"ContainerStarted","Data":"5a2802b9920778b4b100e4bb34bb5ed250972888294087c1098b5a34077d9363"} Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.277742 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-qsqkp" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.279886 4808 patch_prober.go:28] interesting pod/downloads-7954f5f757-qsqkp container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.280450 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qsqkp" podUID="8bc8bd65-c754-4a0a-9f11-40d444e2492a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.283154 4808 generic.go:334] "Generic (PLEG): container finished" podID="0ce081a7-e90f-40ac-aa28-5ef6e19af29a" containerID="09ab13f0fec80e081e3b6892c80040b36447722611c93b243691dbdf384ff48c" exitCode=0 Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.283249 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lbw7q" event={"ID":"0ce081a7-e90f-40ac-aa28-5ef6e19af29a","Type":"ContainerDied","Data":"09ab13f0fec80e081e3b6892c80040b36447722611c93b243691dbdf384ff48c"} Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.283285 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lbw7q" event={"ID":"0ce081a7-e90f-40ac-aa28-5ef6e19af29a","Type":"ContainerStarted","Data":"ed1ed0cab80c5806ed19c6a8eb0dd583154d244ed8733da4f35cd1fe66273f8a"} Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.285425 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-t4qrz" event={"ID":"23bd8519-5b14-499f-9277-154d7397fed6","Type":"ContainerStarted","Data":"17bd6d5aff9caf8afca18f83b0abe7b4420d7658ad500a02d5da55e4bc4e2f32"} Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.285447 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-t4qrz" event={"ID":"23bd8519-5b14-499f-9277-154d7397fed6","Type":"ContainerStarted","Data":"bcb551a0f1a2760aa9dd587d06fe438085ee6ffa3052f2d5c02f04b402a9d16c"} Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.286195 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-t4qrz" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.291867 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bf03d573-aaae-4b60-b285-da3bb7d3e4e3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zmglv\" (UID: \"bf03d573-aaae-4b60-b285-da3bb7d3e4e3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmglv" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.309899 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.310769 4808 patch_prober.go:28] interesting pod/console-operator-58897d9998-t4qrz container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.310918 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-t4qrz" podUID="23bd8519-5b14-499f-9277-154d7397fed6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.314693 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhjq9\" (UniqueName: \"kubernetes.io/projected/f6a13086-47b7-4624-bd50-773a4badcb3f-kube-api-access-lhjq9\") pod \"migrator-59844c95c7-6vlpf\" (UID: \"f6a13086-47b7-4624-bd50-773a4badcb3f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.318144 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5zh9" event={"ID":"44a4d77b-db5b-41ac-89b2-a2eadab424ac","Type":"ContainerStarted","Data":"9d9d5e6a4cf5299ae792b45722af99b3db6b06945e55943e5d0a0a92d678400c"} Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.318200 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5zh9" event={"ID":"44a4d77b-db5b-41ac-89b2-a2eadab424ac","Type":"ContainerStarted","Data":"5cf12ff9b880e6fab043502e553174090407b5ce0e41ce315e8b0eadbc41cec2"} Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.318217 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5zh9" event={"ID":"44a4d77b-db5b-41ac-89b2-a2eadab424ac","Type":"ContainerStarted","Data":"3003e7a835cba69808edf93c4d96f900e8d1994f140d4e3cb4bec63d903b6ddf"} Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.323426 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-lkxxm" event={"ID":"2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205","Type":"ContainerStarted","Data":"d80321049fec7b0127003afc8be505c54b9559a05fd7c37a4048da72448ce3de"} Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.335815 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9xvr" event={"ID":"08c9791c-01e0-4629-b121-2290fd921ff5","Type":"ContainerStarted","Data":"bc960a169e1d3b4c705e3986d7a7d9c423343732f58ffa493bdbc31b8d176782"} Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.335881 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9xvr" event={"ID":"08c9791c-01e0-4629-b121-2290fd921ff5","Type":"ContainerStarted","Data":"f6f87bbb381b038767b7986f0e281fa48bbfe499e1b27ddc55982632a0ff81f5"} Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.340672 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s78xt\" (UniqueName: \"kubernetes.io/projected/580bbc4a-b873-4631-86c3-476a771b9a70-kube-api-access-s78xt\") pod \"catalog-operator-68c6474976-ppxrn\" (UID: \"580bbc4a-b873-4631-86c3-476a771b9a70\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ppxrn" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.343989 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.356880 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/13d94271-250f-49f2-b1ef-4a97e991a124-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kp2fm\" (UID: \"13d94271-250f-49f2-b1ef-4a97e991a124\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kp2fm" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.370374 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmglv" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.399531 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnxsg\" (UniqueName: \"kubernetes.io/projected/05b1a2b9-4809-4e5b-87e3-49b61d3c5184-kube-api-access-qnxsg\") pod \"ingress-operator-5b745b69d9-zxfsc\" (UID: \"05b1a2b9-4809-4e5b-87e3-49b61d3c5184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.403922 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ffv8w"] Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.407292 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/05b1a2b9-4809-4e5b-87e3-49b61d3c5184-bound-sa-token\") pod \"ingress-operator-5b745b69d9-zxfsc\" (UID: \"05b1a2b9-4809-4e5b-87e3-49b61d3c5184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.410732 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9qhp\" (UniqueName: \"kubernetes.io/projected/7777a7df-de5d-4287-a6e9-bcb70a07fdfd-kube-api-access-z9qhp\") pod \"multus-admission-controller-857f4d67dd-tj972\" (UID: \"7777a7df-de5d-4287-a6e9-bcb70a07fdfd\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tj972" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.419245 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sksnr" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.431763 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-zqnpj" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.432226 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ppxrn" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.436134 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb48v\" (UniqueName: \"kubernetes.io/projected/6ae1cc03-9560-4e85-889e-63eda4518302-kube-api-access-xb48v\") pod \"kube-storage-version-migrator-operator-b67b599dd-2htct\" (UID: \"6ae1cc03-9560-4e85-889e-63eda4518302\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2htct" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.436353 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf" Oct 02 16:43:04 crc kubenswrapper[4808]: W1002 16:43:04.436812 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43468992_0b2c_47ae_8293_609319017a71.slice/crio-3a4bd0f5b6e91f7aefc4d307ac4e4c2b161c74092b3b9ee242d9f37dba868917 WatchSource:0}: Error finding container 3a4bd0f5b6e91f7aefc4d307ac4e4c2b161c74092b3b9ee242d9f37dba868917: Status 404 returned error can't find the container with id 3a4bd0f5b6e91f7aefc4d307ac4e4c2b161c74092b3b9ee242d9f37dba868917 Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.445837 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.453042 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmqj6\" (UniqueName: \"kubernetes.io/projected/9ebb9933-4652-4152-a529-5489fd9d5e5d-kube-api-access-kmqj6\") pod \"router-default-5444994796-7b9zg\" (UID: \"9ebb9933-4652-4152-a529-5489fd9d5e5d\") " pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.458215 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2htct" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.466968 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kp2fm" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.475954 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.512546 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-tj972" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.522387 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.535953 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.536044 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/71aafdd1-5917-4764-9503-4c266977f251-registry-certificates\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.536072 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71aafdd1-5917-4764-9503-4c266977f251-trusted-ca\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.536248 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/71aafdd1-5917-4764-9503-4c266977f251-bound-sa-token\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.536351 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tkzc\" (UniqueName: \"kubernetes.io/projected/2849754c-63a9-481f-8c8b-f328a0daaa2c-kube-api-access-5tkzc\") pod \"control-plane-machine-set-operator-78cbb6b69f-qmfsh\" (UID: \"2849754c-63a9-481f-8c8b-f328a0daaa2c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmfsh" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.536382 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/71aafdd1-5917-4764-9503-4c266977f251-ca-trust-extracted\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.536415 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn6kd\" (UniqueName: \"kubernetes.io/projected/71aafdd1-5917-4764-9503-4c266977f251-kube-api-access-zn6kd\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.536440 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/71aafdd1-5917-4764-9503-4c266977f251-registry-tls\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.536505 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/71aafdd1-5917-4764-9503-4c266977f251-installation-pull-secrets\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.536536 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2849754c-63a9-481f-8c8b-f328a0daaa2c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qmfsh\" (UID: \"2849754c-63a9-481f-8c8b-f328a0daaa2c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmfsh" Oct 02 16:43:04 crc kubenswrapper[4808]: E1002 16:43:04.536895 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:05.036875211 +0000 UTC m=+152.362404401 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.570108 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4dk9g"] Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.639592 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.640033 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn6kd\" (UniqueName: \"kubernetes.io/projected/71aafdd1-5917-4764-9503-4c266977f251-kube-api-access-zn6kd\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.640079 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/483da6c8-be0b-4025-9ccb-defe286e2d0e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-n4nnw\" (UID: \"483da6c8-be0b-4025-9ccb-defe286e2d0e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4nnw" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.640129 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c98e3a0-cf25-4f21-acff-13035424b72f-serving-cert\") pod \"service-ca-operator-777779d784-v4bld\" (UID: \"4c98e3a0-cf25-4f21-acff-13035424b72f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4bld" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.640157 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3c05ec42-ae20-46df-9231-a9a98a435103-metrics-tls\") pod \"dns-operator-744455d44c-ddlqv\" (UID: \"3c05ec42-ae20-46df-9231-a9a98a435103\") " pod="openshift-dns-operator/dns-operator-744455d44c-ddlqv" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.640187 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1-metrics-tls\") pod \"dns-default-vh4gp\" (UID: \"7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1\") " pod="openshift-dns/dns-default-vh4gp" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.640320 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kdgh\" (UniqueName: \"kubernetes.io/projected/c74723a8-81cf-4633-8655-e1460025f1b7-kube-api-access-5kdgh\") pod \"controller-manager-879f6c89f-qc629\" (UID: \"c74723a8-81cf-4633-8655-e1460025f1b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.640343 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c74723a8-81cf-4633-8655-e1460025f1b7-client-ca\") pod \"controller-manager-879f6c89f-qc629\" (UID: \"c74723a8-81cf-4633-8655-e1460025f1b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.640505 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/71aafdd1-5917-4764-9503-4c266977f251-registry-certificates\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.640529 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71aafdd1-5917-4764-9503-4c266977f251-trusted-ca\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.640555 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/483da6c8-be0b-4025-9ccb-defe286e2d0e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-n4nnw\" (UID: \"483da6c8-be0b-4025-9ccb-defe286e2d0e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4nnw" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.640579 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/aea5ac39-3315-4186-9002-e836172b6d34-images\") pod \"machine-config-operator-74547568cd-s6jhz\" (UID: \"aea5ac39-3315-4186-9002-e836172b6d34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.640644 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e7f5669-a998-4474-a5cc-47a075e7658f-serving-cert\") pod \"etcd-operator-b45778765-q7ck8\" (UID: \"8e7f5669-a998-4474-a5cc-47a075e7658f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.640731 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/8ddfee45-b557-4f8b-8702-6896ca1fa9ff-tmpfs\") pod \"packageserver-d55dfcdfc-dp2fq\" (UID: \"8ddfee45-b557-4f8b-8702-6896ca1fa9ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.640753 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/aea5ac39-3315-4186-9002-e836172b6d34-auth-proxy-config\") pod \"machine-config-operator-74547568cd-s6jhz\" (UID: \"aea5ac39-3315-4186-9002-e836172b6d34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.640835 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/aea5ac39-3315-4186-9002-e836172b6d34-proxy-tls\") pod \"machine-config-operator-74547568cd-s6jhz\" (UID: \"aea5ac39-3315-4186-9002-e836172b6d34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.640889 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfpqd\" (UniqueName: \"kubernetes.io/projected/3c05ec42-ae20-46df-9231-a9a98a435103-kube-api-access-dfpqd\") pod \"dns-operator-744455d44c-ddlqv\" (UID: \"3c05ec42-ae20-46df-9231-a9a98a435103\") " pod="openshift-dns-operator/dns-operator-744455d44c-ddlqv" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.640922 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8e7f5669-a998-4474-a5cc-47a075e7658f-etcd-ca\") pod \"etcd-operator-b45778765-q7ck8\" (UID: \"8e7f5669-a998-4474-a5cc-47a075e7658f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.640946 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/681e26e4-4764-43f4-b3f9-d084f9619e5f-csi-data-dir\") pod \"csi-hostpathplugin-2htks\" (UID: \"681e26e4-4764-43f4-b3f9-d084f9619e5f\") " pod="hostpath-provisioner/csi-hostpathplugin-2htks" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.640982 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8e7f5669-a998-4474-a5cc-47a075e7658f-etcd-client\") pod \"etcd-operator-b45778765-q7ck8\" (UID: \"8e7f5669-a998-4474-a5cc-47a075e7658f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641002 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c74723a8-81cf-4633-8655-e1460025f1b7-serving-cert\") pod \"controller-manager-879f6c89f-qc629\" (UID: \"c74723a8-81cf-4633-8655-e1460025f1b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641023 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/71aafdd1-5917-4764-9503-4c266977f251-bound-sa-token\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641046 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cwvp\" (UniqueName: \"kubernetes.io/projected/4d096969-852b-4401-bef0-757915f2cdb5-kube-api-access-5cwvp\") pod \"machine-config-server-bxvk8\" (UID: \"4d096969-852b-4401-bef0-757915f2cdb5\") " pod="openshift-machine-config-operator/machine-config-server-bxvk8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641073 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4zsf\" (UniqueName: \"kubernetes.io/projected/e736d6e6-545b-451a-80d8-9f4e6d83b4bc-kube-api-access-q4zsf\") pod \"olm-operator-6b444d44fb-6dtm9\" (UID: \"e736d6e6-545b-451a-80d8-9f4e6d83b4bc\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641120 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1-config-volume\") pod \"dns-default-vh4gp\" (UID: \"7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1\") " pod="openshift-dns/dns-default-vh4gp" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641146 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tkzc\" (UniqueName: \"kubernetes.io/projected/2849754c-63a9-481f-8c8b-f328a0daaa2c-kube-api-access-5tkzc\") pod \"control-plane-machine-set-operator-78cbb6b69f-qmfsh\" (UID: \"2849754c-63a9-481f-8c8b-f328a0daaa2c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmfsh" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641173 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4d89d791-944c-4afc-b0f9-83d3f7604132-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-r4vs4\" (UID: \"4d89d791-944c-4afc-b0f9-83d3f7604132\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r4vs4" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641200 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8ddfee45-b557-4f8b-8702-6896ca1fa9ff-apiservice-cert\") pod \"packageserver-d55dfcdfc-dp2fq\" (UID: \"8ddfee45-b557-4f8b-8702-6896ca1fa9ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641309 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vn5d\" (UniqueName: \"kubernetes.io/projected/4c98e3a0-cf25-4f21-acff-13035424b72f-kube-api-access-2vn5d\") pod \"service-ca-operator-777779d784-v4bld\" (UID: \"4c98e3a0-cf25-4f21-acff-13035424b72f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4bld" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641336 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/71aafdd1-5917-4764-9503-4c266977f251-ca-trust-extracted\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641355 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af55464e-2244-46e0-a2bc-9211829b2d7d-secret-volume\") pod \"collect-profiles-29323710-kllgw\" (UID: \"af55464e-2244-46e0-a2bc-9211829b2d7d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641385 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm5n5\" (UniqueName: \"kubernetes.io/projected/4d89d791-944c-4afc-b0f9-83d3f7604132-kube-api-access-zm5n5\") pod \"package-server-manager-789f6589d5-r4vs4\" (UID: \"4d89d791-944c-4afc-b0f9-83d3f7604132\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r4vs4" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641417 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4d096969-852b-4401-bef0-757915f2cdb5-node-bootstrap-token\") pod \"machine-config-server-bxvk8\" (UID: \"4d096969-852b-4401-bef0-757915f2cdb5\") " pod="openshift-machine-config-operator/machine-config-server-bxvk8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641444 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c74723a8-81cf-4633-8655-e1460025f1b7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qc629\" (UID: \"c74723a8-81cf-4633-8655-e1460025f1b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641511 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e736d6e6-545b-451a-80d8-9f4e6d83b4bc-srv-cert\") pod \"olm-operator-6b444d44fb-6dtm9\" (UID: \"e736d6e6-545b-451a-80d8-9f4e6d83b4bc\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641538 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/71aafdd1-5917-4764-9503-4c266977f251-registry-tls\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641562 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4d096969-852b-4401-bef0-757915f2cdb5-certs\") pod \"machine-config-server-bxvk8\" (UID: \"4d096969-852b-4401-bef0-757915f2cdb5\") " pod="openshift-machine-config-operator/machine-config-server-bxvk8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641582 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8ddfee45-b557-4f8b-8702-6896ca1fa9ff-webhook-cert\") pod \"packageserver-d55dfcdfc-dp2fq\" (UID: \"8ddfee45-b557-4f8b-8702-6896ca1fa9ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641613 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84ppp\" (UniqueName: \"kubernetes.io/projected/483da6c8-be0b-4025-9ccb-defe286e2d0e-kube-api-access-84ppp\") pod \"cluster-image-registry-operator-dc59b4c8b-n4nnw\" (UID: \"483da6c8-be0b-4025-9ccb-defe286e2d0e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4nnw" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641785 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e736d6e6-545b-451a-80d8-9f4e6d83b4bc-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6dtm9\" (UID: \"e736d6e6-545b-451a-80d8-9f4e6d83b4bc\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641810 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b3771709-b077-4820-bfb3-1ea9986866bb-signing-cabundle\") pod \"service-ca-9c57cc56f-7pnwt\" (UID: \"b3771709-b077-4820-bfb3-1ea9986866bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-7pnwt" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641856 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/71aafdd1-5917-4764-9503-4c266977f251-installation-pull-secrets\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641878 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp4w7\" (UniqueName: \"kubernetes.io/projected/9e67480b-9e3f-433f-95f1-ce1e32298fdc-kube-api-access-zp4w7\") pod \"machine-config-controller-84d6567774-p47w4\" (UID: \"9e67480b-9e3f-433f-95f1-ce1e32298fdc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p47w4" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641900 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/02097ffc-aa46-438f-b11d-c98301f3e617-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gpxn2\" (UID: \"02097ffc-aa46-438f-b11d-c98301f3e617\") " pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.641960 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/02097ffc-aa46-438f-b11d-c98301f3e617-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gpxn2\" (UID: \"02097ffc-aa46-438f-b11d-c98301f3e617\") " pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642005 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c74723a8-81cf-4633-8655-e1460025f1b7-config\") pod \"controller-manager-879f6c89f-qc629\" (UID: \"c74723a8-81cf-4633-8655-e1460025f1b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642029 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2849754c-63a9-481f-8c8b-f328a0daaa2c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qmfsh\" (UID: \"2849754c-63a9-481f-8c8b-f328a0daaa2c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmfsh" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642090 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e7f5669-a998-4474-a5cc-47a075e7658f-config\") pod \"etcd-operator-b45778765-q7ck8\" (UID: \"8e7f5669-a998-4474-a5cc-47a075e7658f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642112 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4ncf\" (UniqueName: \"kubernetes.io/projected/8e7f5669-a998-4474-a5cc-47a075e7658f-kube-api-access-q4ncf\") pod \"etcd-operator-b45778765-q7ck8\" (UID: \"8e7f5669-a998-4474-a5cc-47a075e7658f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642131 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq8jk\" (UniqueName: \"kubernetes.io/projected/ee194b74-bef2-4385-8135-7c25d5de5589-kube-api-access-pq8jk\") pod \"ingress-canary-4ff6b\" (UID: \"ee194b74-bef2-4385-8135-7c25d5de5589\") " pod="openshift-ingress-canary/ingress-canary-4ff6b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642189 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/681e26e4-4764-43f4-b3f9-d084f9619e5f-mountpoint-dir\") pod \"csi-hostpathplugin-2htks\" (UID: \"681e26e4-4764-43f4-b3f9-d084f9619e5f\") " pod="hostpath-provisioner/csi-hostpathplugin-2htks" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642207 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/483da6c8-be0b-4025-9ccb-defe286e2d0e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-n4nnw\" (UID: \"483da6c8-be0b-4025-9ccb-defe286e2d0e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4nnw" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642256 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9e67480b-9e3f-433f-95f1-ce1e32298fdc-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-p47w4\" (UID: \"9e67480b-9e3f-433f-95f1-ce1e32298fdc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p47w4" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642304 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt5fj\" (UniqueName: \"kubernetes.io/projected/681e26e4-4764-43f4-b3f9-d084f9619e5f-kube-api-access-bt5fj\") pod \"csi-hostpathplugin-2htks\" (UID: \"681e26e4-4764-43f4-b3f9-d084f9619e5f\") " pod="hostpath-provisioner/csi-hostpathplugin-2htks" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642377 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/681e26e4-4764-43f4-b3f9-d084f9619e5f-socket-dir\") pod \"csi-hostpathplugin-2htks\" (UID: \"681e26e4-4764-43f4-b3f9-d084f9619e5f\") " pod="hostpath-provisioner/csi-hostpathplugin-2htks" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642397 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b3771709-b077-4820-bfb3-1ea9986866bb-signing-key\") pod \"service-ca-9c57cc56f-7pnwt\" (UID: \"b3771709-b077-4820-bfb3-1ea9986866bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-7pnwt" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642417 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsgck\" (UniqueName: \"kubernetes.io/projected/8ddfee45-b557-4f8b-8702-6896ca1fa9ff-kube-api-access-vsgck\") pod \"packageserver-d55dfcdfc-dp2fq\" (UID: \"8ddfee45-b557-4f8b-8702-6896ca1fa9ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642537 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af55464e-2244-46e0-a2bc-9211829b2d7d-config-volume\") pod \"collect-profiles-29323710-kllgw\" (UID: \"af55464e-2244-46e0-a2bc-9211829b2d7d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642556 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kb8h\" (UniqueName: \"kubernetes.io/projected/aea5ac39-3315-4186-9002-e836172b6d34-kube-api-access-8kb8h\") pod \"machine-config-operator-74547568cd-s6jhz\" (UID: \"aea5ac39-3315-4186-9002-e836172b6d34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642594 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc7rr\" (UniqueName: \"kubernetes.io/projected/02097ffc-aa46-438f-b11d-c98301f3e617-kube-api-access-rc7rr\") pod \"marketplace-operator-79b997595-gpxn2\" (UID: \"02097ffc-aa46-438f-b11d-c98301f3e617\") " pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642630 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c98e3a0-cf25-4f21-acff-13035424b72f-config\") pod \"service-ca-operator-777779d784-v4bld\" (UID: \"4c98e3a0-cf25-4f21-acff-13035424b72f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4bld" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642664 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ee194b74-bef2-4385-8135-7c25d5de5589-cert\") pod \"ingress-canary-4ff6b\" (UID: \"ee194b74-bef2-4385-8135-7c25d5de5589\") " pod="openshift-ingress-canary/ingress-canary-4ff6b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642697 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjv9w\" (UniqueName: \"kubernetes.io/projected/b3771709-b077-4820-bfb3-1ea9986866bb-kube-api-access-xjv9w\") pod \"service-ca-9c57cc56f-7pnwt\" (UID: \"b3771709-b077-4820-bfb3-1ea9986866bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-7pnwt" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642719 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9e67480b-9e3f-433f-95f1-ce1e32298fdc-proxy-tls\") pod \"machine-config-controller-84d6567774-p47w4\" (UID: \"9e67480b-9e3f-433f-95f1-ce1e32298fdc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p47w4" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642738 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/681e26e4-4764-43f4-b3f9-d084f9619e5f-plugins-dir\") pod \"csi-hostpathplugin-2htks\" (UID: \"681e26e4-4764-43f4-b3f9-d084f9619e5f\") " pod="hostpath-provisioner/csi-hostpathplugin-2htks" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642785 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8e7f5669-a998-4474-a5cc-47a075e7658f-etcd-service-ca\") pod \"etcd-operator-b45778765-q7ck8\" (UID: \"8e7f5669-a998-4474-a5cc-47a075e7658f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642805 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/681e26e4-4764-43f4-b3f9-d084f9619e5f-registration-dir\") pod \"csi-hostpathplugin-2htks\" (UID: \"681e26e4-4764-43f4-b3f9-d084f9619e5f\") " pod="hostpath-provisioner/csi-hostpathplugin-2htks" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642825 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdtb8\" (UniqueName: \"kubernetes.io/projected/af55464e-2244-46e0-a2bc-9211829b2d7d-kube-api-access-sdtb8\") pod \"collect-profiles-29323710-kllgw\" (UID: \"af55464e-2244-46e0-a2bc-9211829b2d7d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.642855 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmlzr\" (UniqueName: \"kubernetes.io/projected/7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1-kube-api-access-gmlzr\") pod \"dns-default-vh4gp\" (UID: \"7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1\") " pod="openshift-dns/dns-default-vh4gp" Oct 02 16:43:04 crc kubenswrapper[4808]: E1002 16:43:04.652339 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:05.152290202 +0000 UTC m=+152.477819202 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.654266 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/71aafdd1-5917-4764-9503-4c266977f251-registry-certificates\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.655215 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71aafdd1-5917-4764-9503-4c266977f251-trusted-ca\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.655490 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/71aafdd1-5917-4764-9503-4c266977f251-ca-trust-extracted\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.673981 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/71aafdd1-5917-4764-9503-4c266977f251-registry-tls\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.678347 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2849754c-63a9-481f-8c8b-f328a0daaa2c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qmfsh\" (UID: \"2849754c-63a9-481f-8c8b-f328a0daaa2c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmfsh" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.679991 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/71aafdd1-5917-4764-9503-4c266977f251-installation-pull-secrets\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.683391 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2bgkb"] Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.689195 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn6kd\" (UniqueName: \"kubernetes.io/projected/71aafdd1-5917-4764-9503-4c266977f251-kube-api-access-zn6kd\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.696814 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/71aafdd1-5917-4764-9503-4c266977f251-bound-sa-token\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.707083 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-dlddx"] Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.715921 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tkzc\" (UniqueName: \"kubernetes.io/projected/2849754c-63a9-481f-8c8b-f328a0daaa2c-kube-api-access-5tkzc\") pod \"control-plane-machine-set-operator-78cbb6b69f-qmfsh\" (UID: \"2849754c-63a9-481f-8c8b-f328a0daaa2c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmfsh" Oct 02 16:43:04 crc kubenswrapper[4808]: W1002 16:43:04.723903 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc58d3855_bb32_4bb3_94ae_53c49209655e.slice/crio-78ca84097692742cb9b739f329a421350f052830b92da98d0bb31eb70215430f WatchSource:0}: Error finding container 78ca84097692742cb9b739f329a421350f052830b92da98d0bb31eb70215430f: Status 404 returned error can't find the container with id 78ca84097692742cb9b739f329a421350f052830b92da98d0bb31eb70215430f Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.745910 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.745972 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/483da6c8-be0b-4025-9ccb-defe286e2d0e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-n4nnw\" (UID: \"483da6c8-be0b-4025-9ccb-defe286e2d0e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4nnw" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.745991 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/aea5ac39-3315-4186-9002-e836172b6d34-images\") pod \"machine-config-operator-74547568cd-s6jhz\" (UID: \"aea5ac39-3315-4186-9002-e836172b6d34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.746016 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e7f5669-a998-4474-a5cc-47a075e7658f-serving-cert\") pod \"etcd-operator-b45778765-q7ck8\" (UID: \"8e7f5669-a998-4474-a5cc-47a075e7658f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.746042 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/8ddfee45-b557-4f8b-8702-6896ca1fa9ff-tmpfs\") pod \"packageserver-d55dfcdfc-dp2fq\" (UID: \"8ddfee45-b557-4f8b-8702-6896ca1fa9ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.746063 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/aea5ac39-3315-4186-9002-e836172b6d34-auth-proxy-config\") pod \"machine-config-operator-74547568cd-s6jhz\" (UID: \"aea5ac39-3315-4186-9002-e836172b6d34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.746084 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/aea5ac39-3315-4186-9002-e836172b6d34-proxy-tls\") pod \"machine-config-operator-74547568cd-s6jhz\" (UID: \"aea5ac39-3315-4186-9002-e836172b6d34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.746114 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfpqd\" (UniqueName: \"kubernetes.io/projected/3c05ec42-ae20-46df-9231-a9a98a435103-kube-api-access-dfpqd\") pod \"dns-operator-744455d44c-ddlqv\" (UID: \"3c05ec42-ae20-46df-9231-a9a98a435103\") " pod="openshift-dns-operator/dns-operator-744455d44c-ddlqv" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.746135 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8e7f5669-a998-4474-a5cc-47a075e7658f-etcd-ca\") pod \"etcd-operator-b45778765-q7ck8\" (UID: \"8e7f5669-a998-4474-a5cc-47a075e7658f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.746151 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/681e26e4-4764-43f4-b3f9-d084f9619e5f-csi-data-dir\") pod \"csi-hostpathplugin-2htks\" (UID: \"681e26e4-4764-43f4-b3f9-d084f9619e5f\") " pod="hostpath-provisioner/csi-hostpathplugin-2htks" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.746175 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8e7f5669-a998-4474-a5cc-47a075e7658f-etcd-client\") pod \"etcd-operator-b45778765-q7ck8\" (UID: \"8e7f5669-a998-4474-a5cc-47a075e7658f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.746192 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c74723a8-81cf-4633-8655-e1460025f1b7-serving-cert\") pod \"controller-manager-879f6c89f-qc629\" (UID: \"c74723a8-81cf-4633-8655-e1460025f1b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.746211 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cwvp\" (UniqueName: \"kubernetes.io/projected/4d096969-852b-4401-bef0-757915f2cdb5-kube-api-access-5cwvp\") pod \"machine-config-server-bxvk8\" (UID: \"4d096969-852b-4401-bef0-757915f2cdb5\") " pod="openshift-machine-config-operator/machine-config-server-bxvk8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.746247 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4zsf\" (UniqueName: \"kubernetes.io/projected/e736d6e6-545b-451a-80d8-9f4e6d83b4bc-kube-api-access-q4zsf\") pod \"olm-operator-6b444d44fb-6dtm9\" (UID: \"e736d6e6-545b-451a-80d8-9f4e6d83b4bc\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.746245 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmfsh" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.746267 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1-config-volume\") pod \"dns-default-vh4gp\" (UID: \"7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1\") " pod="openshift-dns/dns-default-vh4gp" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.747131 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1-config-volume\") pod \"dns-default-vh4gp\" (UID: \"7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1\") " pod="openshift-dns/dns-default-vh4gp" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.747183 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4d89d791-944c-4afc-b0f9-83d3f7604132-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-r4vs4\" (UID: \"4d89d791-944c-4afc-b0f9-83d3f7604132\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r4vs4" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.747188 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8e7f5669-a998-4474-a5cc-47a075e7658f-etcd-ca\") pod \"etcd-operator-b45778765-q7ck8\" (UID: \"8e7f5669-a998-4474-a5cc-47a075e7658f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.747210 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8ddfee45-b557-4f8b-8702-6896ca1fa9ff-apiservice-cert\") pod \"packageserver-d55dfcdfc-dp2fq\" (UID: \"8ddfee45-b557-4f8b-8702-6896ca1fa9ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.747268 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vn5d\" (UniqueName: \"kubernetes.io/projected/4c98e3a0-cf25-4f21-acff-13035424b72f-kube-api-access-2vn5d\") pod \"service-ca-operator-777779d784-v4bld\" (UID: \"4c98e3a0-cf25-4f21-acff-13035424b72f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4bld" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.747321 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/681e26e4-4764-43f4-b3f9-d084f9619e5f-csi-data-dir\") pod \"csi-hostpathplugin-2htks\" (UID: \"681e26e4-4764-43f4-b3f9-d084f9619e5f\") " pod="hostpath-provisioner/csi-hostpathplugin-2htks" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.747913 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/aea5ac39-3315-4186-9002-e836172b6d34-images\") pod \"machine-config-operator-74547568cd-s6jhz\" (UID: \"aea5ac39-3315-4186-9002-e836172b6d34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.748211 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af55464e-2244-46e0-a2bc-9211829b2d7d-secret-volume\") pod \"collect-profiles-29323710-kllgw\" (UID: \"af55464e-2244-46e0-a2bc-9211829b2d7d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.748321 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm5n5\" (UniqueName: \"kubernetes.io/projected/4d89d791-944c-4afc-b0f9-83d3f7604132-kube-api-access-zm5n5\") pod \"package-server-manager-789f6589d5-r4vs4\" (UID: \"4d89d791-944c-4afc-b0f9-83d3f7604132\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r4vs4" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.748396 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4d096969-852b-4401-bef0-757915f2cdb5-node-bootstrap-token\") pod \"machine-config-server-bxvk8\" (UID: \"4d096969-852b-4401-bef0-757915f2cdb5\") " pod="openshift-machine-config-operator/machine-config-server-bxvk8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.748465 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c74723a8-81cf-4633-8655-e1460025f1b7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qc629\" (UID: \"c74723a8-81cf-4633-8655-e1460025f1b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.749531 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/8ddfee45-b557-4f8b-8702-6896ca1fa9ff-tmpfs\") pod \"packageserver-d55dfcdfc-dp2fq\" (UID: \"8ddfee45-b557-4f8b-8702-6896ca1fa9ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.758658 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8ddfee45-b557-4f8b-8702-6896ca1fa9ff-apiservice-cert\") pod \"packageserver-d55dfcdfc-dp2fq\" (UID: \"8ddfee45-b557-4f8b-8702-6896ca1fa9ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.759911 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/aea5ac39-3315-4186-9002-e836172b6d34-auth-proxy-config\") pod \"machine-config-operator-74547568cd-s6jhz\" (UID: \"aea5ac39-3315-4186-9002-e836172b6d34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.760399 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8e7f5669-a998-4474-a5cc-47a075e7658f-etcd-client\") pod \"etcd-operator-b45778765-q7ck8\" (UID: \"8e7f5669-a998-4474-a5cc-47a075e7658f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" Oct 02 16:43:04 crc kubenswrapper[4808]: E1002 16:43:04.760813 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:05.260787857 +0000 UTC m=+152.586316857 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.765579 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e736d6e6-545b-451a-80d8-9f4e6d83b4bc-srv-cert\") pod \"olm-operator-6b444d44fb-6dtm9\" (UID: \"e736d6e6-545b-451a-80d8-9f4e6d83b4bc\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.767029 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c74723a8-81cf-4633-8655-e1460025f1b7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qc629\" (UID: \"c74723a8-81cf-4633-8655-e1460025f1b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.770515 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/aea5ac39-3315-4186-9002-e836172b6d34-proxy-tls\") pod \"machine-config-operator-74547568cd-s6jhz\" (UID: \"aea5ac39-3315-4186-9002-e836172b6d34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.770689 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4d096969-852b-4401-bef0-757915f2cdb5-certs\") pod \"machine-config-server-bxvk8\" (UID: \"4d096969-852b-4401-bef0-757915f2cdb5\") " pod="openshift-machine-config-operator/machine-config-server-bxvk8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.770819 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4d89d791-944c-4afc-b0f9-83d3f7604132-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-r4vs4\" (UID: \"4d89d791-944c-4afc-b0f9-83d3f7604132\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r4vs4" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.770903 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84ppp\" (UniqueName: \"kubernetes.io/projected/483da6c8-be0b-4025-9ccb-defe286e2d0e-kube-api-access-84ppp\") pod \"cluster-image-registry-operator-dc59b4c8b-n4nnw\" (UID: \"483da6c8-be0b-4025-9ccb-defe286e2d0e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4nnw" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.770942 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e736d6e6-545b-451a-80d8-9f4e6d83b4bc-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6dtm9\" (UID: \"e736d6e6-545b-451a-80d8-9f4e6d83b4bc\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.770977 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e7f5669-a998-4474-a5cc-47a075e7658f-serving-cert\") pod \"etcd-operator-b45778765-q7ck8\" (UID: \"8e7f5669-a998-4474-a5cc-47a075e7658f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.770991 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8ddfee45-b557-4f8b-8702-6896ca1fa9ff-webhook-cert\") pod \"packageserver-d55dfcdfc-dp2fq\" (UID: \"8ddfee45-b557-4f8b-8702-6896ca1fa9ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771018 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b3771709-b077-4820-bfb3-1ea9986866bb-signing-cabundle\") pod \"service-ca-9c57cc56f-7pnwt\" (UID: \"b3771709-b077-4820-bfb3-1ea9986866bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-7pnwt" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771062 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/02097ffc-aa46-438f-b11d-c98301f3e617-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gpxn2\" (UID: \"02097ffc-aa46-438f-b11d-c98301f3e617\") " pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771086 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp4w7\" (UniqueName: \"kubernetes.io/projected/9e67480b-9e3f-433f-95f1-ce1e32298fdc-kube-api-access-zp4w7\") pod \"machine-config-controller-84d6567774-p47w4\" (UID: \"9e67480b-9e3f-433f-95f1-ce1e32298fdc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p47w4" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771105 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/02097ffc-aa46-438f-b11d-c98301f3e617-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gpxn2\" (UID: \"02097ffc-aa46-438f-b11d-c98301f3e617\") " pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771154 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c74723a8-81cf-4633-8655-e1460025f1b7-config\") pod \"controller-manager-879f6c89f-qc629\" (UID: \"c74723a8-81cf-4633-8655-e1460025f1b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771227 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e7f5669-a998-4474-a5cc-47a075e7658f-config\") pod \"etcd-operator-b45778765-q7ck8\" (UID: \"8e7f5669-a998-4474-a5cc-47a075e7658f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771312 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4ncf\" (UniqueName: \"kubernetes.io/projected/8e7f5669-a998-4474-a5cc-47a075e7658f-kube-api-access-q4ncf\") pod \"etcd-operator-b45778765-q7ck8\" (UID: \"8e7f5669-a998-4474-a5cc-47a075e7658f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771370 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq8jk\" (UniqueName: \"kubernetes.io/projected/ee194b74-bef2-4385-8135-7c25d5de5589-kube-api-access-pq8jk\") pod \"ingress-canary-4ff6b\" (UID: \"ee194b74-bef2-4385-8135-7c25d5de5589\") " pod="openshift-ingress-canary/ingress-canary-4ff6b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771532 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/681e26e4-4764-43f4-b3f9-d084f9619e5f-mountpoint-dir\") pod \"csi-hostpathplugin-2htks\" (UID: \"681e26e4-4764-43f4-b3f9-d084f9619e5f\") " pod="hostpath-provisioner/csi-hostpathplugin-2htks" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771569 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/483da6c8-be0b-4025-9ccb-defe286e2d0e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-n4nnw\" (UID: \"483da6c8-be0b-4025-9ccb-defe286e2d0e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4nnw" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771613 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9e67480b-9e3f-433f-95f1-ce1e32298fdc-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-p47w4\" (UID: \"9e67480b-9e3f-433f-95f1-ce1e32298fdc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p47w4" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771638 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt5fj\" (UniqueName: \"kubernetes.io/projected/681e26e4-4764-43f4-b3f9-d084f9619e5f-kube-api-access-bt5fj\") pod \"csi-hostpathplugin-2htks\" (UID: \"681e26e4-4764-43f4-b3f9-d084f9619e5f\") " pod="hostpath-provisioner/csi-hostpathplugin-2htks" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771664 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/681e26e4-4764-43f4-b3f9-d084f9619e5f-socket-dir\") pod \"csi-hostpathplugin-2htks\" (UID: \"681e26e4-4764-43f4-b3f9-d084f9619e5f\") " pod="hostpath-provisioner/csi-hostpathplugin-2htks" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771708 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b3771709-b077-4820-bfb3-1ea9986866bb-signing-key\") pod \"service-ca-9c57cc56f-7pnwt\" (UID: \"b3771709-b077-4820-bfb3-1ea9986866bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-7pnwt" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771727 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsgck\" (UniqueName: \"kubernetes.io/projected/8ddfee45-b557-4f8b-8702-6896ca1fa9ff-kube-api-access-vsgck\") pod \"packageserver-d55dfcdfc-dp2fq\" (UID: \"8ddfee45-b557-4f8b-8702-6896ca1fa9ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771784 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af55464e-2244-46e0-a2bc-9211829b2d7d-config-volume\") pod \"collect-profiles-29323710-kllgw\" (UID: \"af55464e-2244-46e0-a2bc-9211829b2d7d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771806 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kb8h\" (UniqueName: \"kubernetes.io/projected/aea5ac39-3315-4186-9002-e836172b6d34-kube-api-access-8kb8h\") pod \"machine-config-operator-74547568cd-s6jhz\" (UID: \"aea5ac39-3315-4186-9002-e836172b6d34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771845 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc7rr\" (UniqueName: \"kubernetes.io/projected/02097ffc-aa46-438f-b11d-c98301f3e617-kube-api-access-rc7rr\") pod \"marketplace-operator-79b997595-gpxn2\" (UID: \"02097ffc-aa46-438f-b11d-c98301f3e617\") " pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771871 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c98e3a0-cf25-4f21-acff-13035424b72f-config\") pod \"service-ca-operator-777779d784-v4bld\" (UID: \"4c98e3a0-cf25-4f21-acff-13035424b72f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4bld" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771891 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ee194b74-bef2-4385-8135-7c25d5de5589-cert\") pod \"ingress-canary-4ff6b\" (UID: \"ee194b74-bef2-4385-8135-7c25d5de5589\") " pod="openshift-ingress-canary/ingress-canary-4ff6b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.771965 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjv9w\" (UniqueName: \"kubernetes.io/projected/b3771709-b077-4820-bfb3-1ea9986866bb-kube-api-access-xjv9w\") pod \"service-ca-9c57cc56f-7pnwt\" (UID: \"b3771709-b077-4820-bfb3-1ea9986866bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-7pnwt" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.772010 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/681e26e4-4764-43f4-b3f9-d084f9619e5f-plugins-dir\") pod \"csi-hostpathplugin-2htks\" (UID: \"681e26e4-4764-43f4-b3f9-d084f9619e5f\") " pod="hostpath-provisioner/csi-hostpathplugin-2htks" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.772035 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9e67480b-9e3f-433f-95f1-ce1e32298fdc-proxy-tls\") pod \"machine-config-controller-84d6567774-p47w4\" (UID: \"9e67480b-9e3f-433f-95f1-ce1e32298fdc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p47w4" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.772058 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8e7f5669-a998-4474-a5cc-47a075e7658f-etcd-service-ca\") pod \"etcd-operator-b45778765-q7ck8\" (UID: \"8e7f5669-a998-4474-a5cc-47a075e7658f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.772095 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/681e26e4-4764-43f4-b3f9-d084f9619e5f-registration-dir\") pod \"csi-hostpathplugin-2htks\" (UID: \"681e26e4-4764-43f4-b3f9-d084f9619e5f\") " pod="hostpath-provisioner/csi-hostpathplugin-2htks" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.772119 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdtb8\" (UniqueName: \"kubernetes.io/projected/af55464e-2244-46e0-a2bc-9211829b2d7d-kube-api-access-sdtb8\") pod \"collect-profiles-29323710-kllgw\" (UID: \"af55464e-2244-46e0-a2bc-9211829b2d7d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.772140 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmlzr\" (UniqueName: \"kubernetes.io/projected/7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1-kube-api-access-gmlzr\") pod \"dns-default-vh4gp\" (UID: \"7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1\") " pod="openshift-dns/dns-default-vh4gp" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.772183 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/483da6c8-be0b-4025-9ccb-defe286e2d0e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-n4nnw\" (UID: \"483da6c8-be0b-4025-9ccb-defe286e2d0e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4nnw" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.772203 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3c05ec42-ae20-46df-9231-a9a98a435103-metrics-tls\") pod \"dns-operator-744455d44c-ddlqv\" (UID: \"3c05ec42-ae20-46df-9231-a9a98a435103\") " pod="openshift-dns-operator/dns-operator-744455d44c-ddlqv" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.772219 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c98e3a0-cf25-4f21-acff-13035424b72f-serving-cert\") pod \"service-ca-operator-777779d784-v4bld\" (UID: \"4c98e3a0-cf25-4f21-acff-13035424b72f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4bld" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.772304 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1-metrics-tls\") pod \"dns-default-vh4gp\" (UID: \"7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1\") " pod="openshift-dns/dns-default-vh4gp" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.772365 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kdgh\" (UniqueName: \"kubernetes.io/projected/c74723a8-81cf-4633-8655-e1460025f1b7-kube-api-access-5kdgh\") pod \"controller-manager-879f6c89f-qc629\" (UID: \"c74723a8-81cf-4633-8655-e1460025f1b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.772398 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c74723a8-81cf-4633-8655-e1460025f1b7-client-ca\") pod \"controller-manager-879f6c89f-qc629\" (UID: \"c74723a8-81cf-4633-8655-e1460025f1b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.774329 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8e7f5669-a998-4474-a5cc-47a075e7658f-etcd-service-ca\") pod \"etcd-operator-b45778765-q7ck8\" (UID: \"8e7f5669-a998-4474-a5cc-47a075e7658f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.774702 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/681e26e4-4764-43f4-b3f9-d084f9619e5f-registration-dir\") pod \"csi-hostpathplugin-2htks\" (UID: \"681e26e4-4764-43f4-b3f9-d084f9619e5f\") " pod="hostpath-provisioner/csi-hostpathplugin-2htks" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.775220 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c98e3a0-cf25-4f21-acff-13035424b72f-config\") pod \"service-ca-operator-777779d784-v4bld\" (UID: \"4c98e3a0-cf25-4f21-acff-13035424b72f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4bld" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.776004 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b3771709-b077-4820-bfb3-1ea9986866bb-signing-cabundle\") pod \"service-ca-9c57cc56f-7pnwt\" (UID: \"b3771709-b077-4820-bfb3-1ea9986866bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-7pnwt" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.776084 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/681e26e4-4764-43f4-b3f9-d084f9619e5f-plugins-dir\") pod \"csi-hostpathplugin-2htks\" (UID: \"681e26e4-4764-43f4-b3f9-d084f9619e5f\") " pod="hostpath-provisioner/csi-hostpathplugin-2htks" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.778251 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c74723a8-81cf-4633-8655-e1460025f1b7-client-ca\") pod \"controller-manager-879f6c89f-qc629\" (UID: \"c74723a8-81cf-4633-8655-e1460025f1b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.780288 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c98e3a0-cf25-4f21-acff-13035424b72f-serving-cert\") pod \"service-ca-operator-777779d784-v4bld\" (UID: \"4c98e3a0-cf25-4f21-acff-13035424b72f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4bld" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.782946 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3c05ec42-ae20-46df-9231-a9a98a435103-metrics-tls\") pod \"dns-operator-744455d44c-ddlqv\" (UID: \"3c05ec42-ae20-46df-9231-a9a98a435103\") " pod="openshift-dns-operator/dns-operator-744455d44c-ddlqv" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.782974 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8ddfee45-b557-4f8b-8702-6896ca1fa9ff-webhook-cert\") pod \"packageserver-d55dfcdfc-dp2fq\" (UID: \"8ddfee45-b557-4f8b-8702-6896ca1fa9ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.782931 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e736d6e6-545b-451a-80d8-9f4e6d83b4bc-srv-cert\") pod \"olm-operator-6b444d44fb-6dtm9\" (UID: \"e736d6e6-545b-451a-80d8-9f4e6d83b4bc\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.784318 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e7f5669-a998-4474-a5cc-47a075e7658f-config\") pod \"etcd-operator-b45778765-q7ck8\" (UID: \"8e7f5669-a998-4474-a5cc-47a075e7658f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.784496 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/681e26e4-4764-43f4-b3f9-d084f9619e5f-socket-dir\") pod \"csi-hostpathplugin-2htks\" (UID: \"681e26e4-4764-43f4-b3f9-d084f9619e5f\") " pod="hostpath-provisioner/csi-hostpathplugin-2htks" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.785004 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e736d6e6-545b-451a-80d8-9f4e6d83b4bc-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6dtm9\" (UID: \"e736d6e6-545b-451a-80d8-9f4e6d83b4bc\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.785257 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/02097ffc-aa46-438f-b11d-c98301f3e617-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gpxn2\" (UID: \"02097ffc-aa46-438f-b11d-c98301f3e617\") " pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.785686 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/681e26e4-4764-43f4-b3f9-d084f9619e5f-mountpoint-dir\") pod \"csi-hostpathplugin-2htks\" (UID: \"681e26e4-4764-43f4-b3f9-d084f9619e5f\") " pod="hostpath-provisioner/csi-hostpathplugin-2htks" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.786489 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9e67480b-9e3f-433f-95f1-ce1e32298fdc-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-p47w4\" (UID: \"9e67480b-9e3f-433f-95f1-ce1e32298fdc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p47w4" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.789294 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af55464e-2244-46e0-a2bc-9211829b2d7d-config-volume\") pod \"collect-profiles-29323710-kllgw\" (UID: \"af55464e-2244-46e0-a2bc-9211829b2d7d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.789435 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b3771709-b077-4820-bfb3-1ea9986866bb-signing-key\") pod \"service-ca-9c57cc56f-7pnwt\" (UID: \"b3771709-b077-4820-bfb3-1ea9986866bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-7pnwt" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.789483 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4d096969-852b-4401-bef0-757915f2cdb5-certs\") pod \"machine-config-server-bxvk8\" (UID: \"4d096969-852b-4401-bef0-757915f2cdb5\") " pod="openshift-machine-config-operator/machine-config-server-bxvk8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.791006 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/483da6c8-be0b-4025-9ccb-defe286e2d0e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-n4nnw\" (UID: \"483da6c8-be0b-4025-9ccb-defe286e2d0e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4nnw" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.798997 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c74723a8-81cf-4633-8655-e1460025f1b7-serving-cert\") pod \"controller-manager-879f6c89f-qc629\" (UID: \"c74723a8-81cf-4633-8655-e1460025f1b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.799066 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4d096969-852b-4401-bef0-757915f2cdb5-node-bootstrap-token\") pod \"machine-config-server-bxvk8\" (UID: \"4d096969-852b-4401-bef0-757915f2cdb5\") " pod="openshift-machine-config-operator/machine-config-server-bxvk8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.799186 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1-metrics-tls\") pod \"dns-default-vh4gp\" (UID: \"7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1\") " pod="openshift-dns/dns-default-vh4gp" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.799564 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/02097ffc-aa46-438f-b11d-c98301f3e617-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gpxn2\" (UID: \"02097ffc-aa46-438f-b11d-c98301f3e617\") " pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.799861 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/483da6c8-be0b-4025-9ccb-defe286e2d0e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-n4nnw\" (UID: \"483da6c8-be0b-4025-9ccb-defe286e2d0e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4nnw" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.800150 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ee194b74-bef2-4385-8135-7c25d5de5589-cert\") pod \"ingress-canary-4ff6b\" (UID: \"ee194b74-bef2-4385-8135-7c25d5de5589\") " pod="openshift-ingress-canary/ingress-canary-4ff6b" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.800181 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4zsf\" (UniqueName: \"kubernetes.io/projected/e736d6e6-545b-451a-80d8-9f4e6d83b4bc-kube-api-access-q4zsf\") pod \"olm-operator-6b444d44fb-6dtm9\" (UID: \"e736d6e6-545b-451a-80d8-9f4e6d83b4bc\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.800806 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af55464e-2244-46e0-a2bc-9211829b2d7d-secret-volume\") pod \"collect-profiles-29323710-kllgw\" (UID: \"af55464e-2244-46e0-a2bc-9211829b2d7d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.805301 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c74723a8-81cf-4633-8655-e1460025f1b7-config\") pod \"controller-manager-879f6c89f-qc629\" (UID: \"c74723a8-81cf-4633-8655-e1460025f1b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.803597 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8974h"] Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.809743 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9e67480b-9e3f-433f-95f1-ce1e32298fdc-proxy-tls\") pod \"machine-config-controller-84d6567774-p47w4\" (UID: \"9e67480b-9e3f-433f-95f1-ce1e32298fdc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p47w4" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.840073 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmglv"] Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.842808 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vn5d\" (UniqueName: \"kubernetes.io/projected/4c98e3a0-cf25-4f21-acff-13035424b72f-kube-api-access-2vn5d\") pod \"service-ca-operator-777779d784-v4bld\" (UID: \"4c98e3a0-cf25-4f21-acff-13035424b72f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4bld" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.847660 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4bld" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.880491 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:04 crc kubenswrapper[4808]: E1002 16:43:04.881024 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:05.381002308 +0000 UTC m=+152.706531308 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.884179 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cwvp\" (UniqueName: \"kubernetes.io/projected/4d096969-852b-4401-bef0-757915f2cdb5-kube-api-access-5cwvp\") pod \"machine-config-server-bxvk8\" (UID: \"4d096969-852b-4401-bef0-757915f2cdb5\") " pod="openshift-machine-config-operator/machine-config-server-bxvk8" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.884337 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84ppp\" (UniqueName: \"kubernetes.io/projected/483da6c8-be0b-4025-9ccb-defe286e2d0e-kube-api-access-84ppp\") pod \"cluster-image-registry-operator-dc59b4c8b-n4nnw\" (UID: \"483da6c8-be0b-4025-9ccb-defe286e2d0e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4nnw" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.885963 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfpqd\" (UniqueName: \"kubernetes.io/projected/3c05ec42-ae20-46df-9231-a9a98a435103-kube-api-access-dfpqd\") pod \"dns-operator-744455d44c-ddlqv\" (UID: \"3c05ec42-ae20-46df-9231-a9a98a435103\") " pod="openshift-dns-operator/dns-operator-744455d44c-ddlqv" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.886929 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf"] Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.895771 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm5n5\" (UniqueName: \"kubernetes.io/projected/4d89d791-944c-4afc-b0f9-83d3f7604132-kube-api-access-zm5n5\") pod \"package-server-manager-789f6589d5-r4vs4\" (UID: \"4d89d791-944c-4afc-b0f9-83d3f7604132\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r4vs4" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.922573 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp4w7\" (UniqueName: \"kubernetes.io/projected/9e67480b-9e3f-433f-95f1-ce1e32298fdc-kube-api-access-zp4w7\") pod \"machine-config-controller-84d6567774-p47w4\" (UID: \"9e67480b-9e3f-433f-95f1-ce1e32298fdc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p47w4" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.937507 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-ddlqv" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.947022 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r4vs4" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.952096 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/483da6c8-be0b-4025-9ccb-defe286e2d0e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-n4nnw\" (UID: \"483da6c8-be0b-4025-9ccb-defe286e2d0e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4nnw" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.958670 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.960379 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kb8h\" (UniqueName: \"kubernetes.io/projected/aea5ac39-3315-4186-9002-e836172b6d34-kube-api-access-8kb8h\") pod \"machine-config-operator-74547568cd-s6jhz\" (UID: \"aea5ac39-3315-4186-9002-e836172b6d34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz" Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.975049 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc7rr\" (UniqueName: \"kubernetes.io/projected/02097ffc-aa46-438f-b11d-c98301f3e617-kube-api-access-rc7rr\") pod \"marketplace-operator-79b997595-gpxn2\" (UID: \"02097ffc-aa46-438f-b11d-c98301f3e617\") " pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" Oct 02 16:43:04 crc kubenswrapper[4808]: W1002 16:43:04.976715 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode615c64c_94fd_44fa_a119_69435c37bd82.slice/crio-cdd30bb604a1a73a7938e55aef8bdf55e586377937299fb250277dea03379e1b WatchSource:0}: Error finding container cdd30bb604a1a73a7938e55aef8bdf55e586377937299fb250277dea03379e1b: Status 404 returned error can't find the container with id cdd30bb604a1a73a7938e55aef8bdf55e586377937299fb250277dea03379e1b Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.982192 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:04 crc kubenswrapper[4808]: E1002 16:43:04.982613 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:05.482597961 +0000 UTC m=+152.808126961 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:04 crc kubenswrapper[4808]: I1002 16:43:04.997729 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdtb8\" (UniqueName: \"kubernetes.io/projected/af55464e-2244-46e0-a2bc-9211829b2d7d-kube-api-access-sdtb8\") pod \"collect-profiles-29323710-kllgw\" (UID: \"af55464e-2244-46e0-a2bc-9211829b2d7d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.006916 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.026761 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmlzr\" (UniqueName: \"kubernetes.io/projected/7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1-kube-api-access-gmlzr\") pod \"dns-default-vh4gp\" (UID: \"7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1\") " pod="openshift-dns/dns-default-vh4gp" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.031951 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjv9w\" (UniqueName: \"kubernetes.io/projected/b3771709-b077-4820-bfb3-1ea9986866bb-kube-api-access-xjv9w\") pod \"service-ca-9c57cc56f-7pnwt\" (UID: \"b3771709-b077-4820-bfb3-1ea9986866bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-7pnwt" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.041767 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bxvk8" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.060447 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-vh4gp" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.079165 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kdgh\" (UniqueName: \"kubernetes.io/projected/c74723a8-81cf-4633-8655-e1460025f1b7-kube-api-access-5kdgh\") pod \"controller-manager-879f6c89f-qc629\" (UID: \"c74723a8-81cf-4633-8655-e1460025f1b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.084083 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:05 crc kubenswrapper[4808]: E1002 16:43:05.084499 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:05.584478581 +0000 UTC m=+152.910007581 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.093979 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4ncf\" (UniqueName: \"kubernetes.io/projected/8e7f5669-a998-4474-a5cc-47a075e7658f-kube-api-access-q4ncf\") pod \"etcd-operator-b45778765-q7ck8\" (UID: \"8e7f5669-a998-4474-a5cc-47a075e7658f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.119396 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq8jk\" (UniqueName: \"kubernetes.io/projected/ee194b74-bef2-4385-8135-7c25d5de5589-kube-api-access-pq8jk\") pod \"ingress-canary-4ff6b\" (UID: \"ee194b74-bef2-4385-8135-7c25d5de5589\") " pod="openshift-ingress-canary/ingress-canary-4ff6b" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.131194 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsgck\" (UniqueName: \"kubernetes.io/projected/8ddfee45-b557-4f8b-8702-6896ca1fa9ff-kube-api-access-vsgck\") pod \"packageserver-d55dfcdfc-dp2fq\" (UID: \"8ddfee45-b557-4f8b-8702-6896ca1fa9ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.131929 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt5fj\" (UniqueName: \"kubernetes.io/projected/681e26e4-4764-43f4-b3f9-d084f9619e5f-kube-api-access-bt5fj\") pod \"csi-hostpathplugin-2htks\" (UID: \"681e26e4-4764-43f4-b3f9-d084f9619e5f\") " pod="hostpath-provisioner/csi-hostpathplugin-2htks" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.161276 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p47w4" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.174930 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.186223 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:05 crc kubenswrapper[4808]: E1002 16:43:05.186782 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:05.686765461 +0000 UTC m=+153.012294461 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.200362 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4nnw" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.205182 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.220971 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.263451 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ppxrn"] Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.272017 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.280088 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.287648 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:05 crc kubenswrapper[4808]: E1002 16:43:05.287839 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:05.78780728 +0000 UTC m=+153.113336280 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.287933 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:05 crc kubenswrapper[4808]: E1002 16:43:05.288440 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:05.788423296 +0000 UTC m=+153.113952296 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.289522 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-7pnwt" Oct 02 16:43:05 crc kubenswrapper[4808]: W1002 16:43:05.292463 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d096969_852b_4401_bef0_757915f2cdb5.slice/crio-213a0f86cbb07d04ccf84fada461058df8c14ea60d722185a4950a47e96398db WatchSource:0}: Error finding container 213a0f86cbb07d04ccf84fada461058df8c14ea60d722185a4950a47e96398db: Status 404 returned error can't find the container with id 213a0f86cbb07d04ccf84fada461058df8c14ea60d722185a4950a47e96398db Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.315597 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-4ff6b" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.331976 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2htks" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.373821 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-dlddx" event={"ID":"c58d3855-bb32-4bb3-94ae-53c49209655e","Type":"ContainerStarted","Data":"747c3304769acdbfac095d6353f25b73ea9de1393a4aea5f785e491d56d1445c"} Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.373873 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-dlddx" event={"ID":"c58d3855-bb32-4bb3-94ae-53c49209655e","Type":"ContainerStarted","Data":"78ca84097692742cb9b739f329a421350f052830b92da98d0bb31eb70215430f"} Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.389626 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:05 crc kubenswrapper[4808]: E1002 16:43:05.390368 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:05.890344827 +0000 UTC m=+153.215873827 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.476792 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9xvr" podStartSLOduration=131.476763209 podStartE2EDuration="2m11.476763209s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:05.474484911 +0000 UTC m=+152.800013911" watchObservedRunningTime="2025-10-02 16:43:05.476763209 +0000 UTC m=+152.802292209" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.492935 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zqnpj"] Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.493758 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf"] Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.493792 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4dk9g" event={"ID":"1856b6cf-f125-4641-a31f-7795b811e2c0","Type":"ContainerStarted","Data":"9d7e9dda8cae23b3810adc30d019750c0e691d8716ea50e4ff4c2a57f8745d58"} Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.493834 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4dk9g" event={"ID":"1856b6cf-f125-4641-a31f-7795b811e2c0","Type":"ContainerStarted","Data":"370052ecbf517578fefdc05e54e4b473ffdc49a35e4935107c34a1c716fec7b8"} Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.493850 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" event={"ID":"67d706bc-9776-4edf-98e3-7eb03ff5ef75","Type":"ContainerStarted","Data":"71c748711eafe49584b07368a42d3fafe166b776c0aa38842ce70ba1cae17905"} Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.494380 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2htct"] Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.500125 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sksnr"] Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.496789 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.500573 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-7b9zg" event={"ID":"9ebb9933-4652-4152-a529-5489fd9d5e5d","Type":"ContainerStarted","Data":"0765ff3aaa8ced9f7572317e447126bd23b0c4904145f4720e6df40a0bb144d8"} Oct 02 16:43:05 crc kubenswrapper[4808]: E1002 16:43:05.506626 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:06.006594548 +0000 UTC m=+153.332123548 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.519998 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ffv8w" event={"ID":"43468992-0b2c-47ae-8293-609319017a71","Type":"ContainerStarted","Data":"3df4078ba60bc1a0abc0f14576b86e9168da7e6d344ec4dec3ce360aad73e192"} Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.520059 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ffv8w" event={"ID":"43468992-0b2c-47ae-8293-609319017a71","Type":"ContainerStarted","Data":"3a4bd0f5b6e91f7aefc4d307ac4e4c2b161c74092b3b9ee242d9f37dba868917"} Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.561638 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lbw7q" event={"ID":"0ce081a7-e90f-40ac-aa28-5ef6e19af29a","Type":"ContainerStarted","Data":"cdce6a743b63af82aa5002265bd316da8f5dec9f6a92f5d672d2f22e6d26d321"} Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.561787 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lbw7q" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.573807 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bxvk8" event={"ID":"4d096969-852b-4401-bef0-757915f2cdb5","Type":"ContainerStarted","Data":"213a0f86cbb07d04ccf84fada461058df8c14ea60d722185a4950a47e96398db"} Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.581028 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmglv" event={"ID":"bf03d573-aaae-4b60-b285-da3bb7d3e4e3","Type":"ContainerStarted","Data":"16f97858fbdc1215d34471fef6612a388d3b5a1f207753758b4133ccaaa7d60c"} Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.582933 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8974h" event={"ID":"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324","Type":"ContainerStarted","Data":"82bce0aaceee20db1d78fbc7371d275ab6a03fcafeaab7dc7cd77d300798735d"} Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.585760 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" event={"ID":"e615c64c-94fd-44fa-a119-69435c37bd82","Type":"ContainerStarted","Data":"cdd30bb604a1a73a7938e55aef8bdf55e586377937299fb250277dea03379e1b"} Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.589908 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-lkxxm" event={"ID":"2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205","Type":"ContainerStarted","Data":"0051625984b74d2de17037997cb649dde7854f8f59a6c5a71e1ff58c87b512af"} Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.592941 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ppxrn" event={"ID":"580bbc4a-b873-4631-86c3-476a771b9a70","Type":"ContainerStarted","Data":"516125d7204ddaa330ef53d8e3bdd2e49fb6bacd22c212c8c31124a6bc942d3f"} Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.597787 4808 patch_prober.go:28] interesting pod/downloads-7954f5f757-qsqkp container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.597863 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qsqkp" podUID="8bc8bd65-c754-4a0a-9f11-40d444e2492a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.602804 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:05 crc kubenswrapper[4808]: E1002 16:43:05.603112 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:06.103067672 +0000 UTC m=+153.428596672 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.603457 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.603897 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-qsqkp" podStartSLOduration=131.603874663 podStartE2EDuration="2m11.603874663s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:05.57195153 +0000 UTC m=+152.897480531" watchObservedRunningTime="2025-10-02 16:43:05.603874663 +0000 UTC m=+152.929403663" Oct 02 16:43:05 crc kubenswrapper[4808]: E1002 16:43:05.607086 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:06.107064043 +0000 UTC m=+153.432593043 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.708891 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:05 crc kubenswrapper[4808]: E1002 16:43:05.710754 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:06.210727758 +0000 UTC m=+153.536256758 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.812702 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.812688 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5zh9" podStartSLOduration=131.812661509 podStartE2EDuration="2m11.812661509s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:05.811959082 +0000 UTC m=+153.137488102" watchObservedRunningTime="2025-10-02 16:43:05.812661509 +0000 UTC m=+153.138190509" Oct 02 16:43:05 crc kubenswrapper[4808]: E1002 16:43:05.813037 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:06.313022508 +0000 UTC m=+153.638551498 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.854887 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-t4qrz" podStartSLOduration=131.854856019 podStartE2EDuration="2m11.854856019s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:05.853652749 +0000 UTC m=+153.179181749" watchObservedRunningTime="2025-10-02 16:43:05.854856019 +0000 UTC m=+153.180385019" Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.916211 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:05 crc kubenswrapper[4808]: E1002 16:43:05.916403 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:06.416367945 +0000 UTC m=+153.741896945 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.916542 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:05 crc kubenswrapper[4808]: E1002 16:43:05.916866 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:06.416853447 +0000 UTC m=+153.742382447 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:05 crc kubenswrapper[4808]: I1002 16:43:05.942060 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-t4qrz" Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.019798 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:06 crc kubenswrapper[4808]: E1002 16:43:06.020844 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:06.52082156 +0000 UTC m=+153.846350570 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.122916 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:06 crc kubenswrapper[4808]: E1002 16:43:06.146598 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:06.646574329 +0000 UTC m=+153.972103329 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.169356 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc"] Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.188292 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-tj972"] Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.202332 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kp2fm"] Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.224114 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:06 crc kubenswrapper[4808]: E1002 16:43:06.224698 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:06.724675842 +0000 UTC m=+154.050204842 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.269823 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m"] Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.327896 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:06 crc kubenswrapper[4808]: E1002 16:43:06.328022 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:06.828002548 +0000 UTC m=+154.153531548 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.432452 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:06 crc kubenswrapper[4808]: E1002 16:43:06.432871 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:06.932854923 +0000 UTC m=+154.258383923 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.478497 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.485965 4808 patch_prober.go:28] interesting pod/router-default-5444994796-7b9zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 16:43:06 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Oct 02 16:43:06 crc kubenswrapper[4808]: [+]process-running ok Oct 02 16:43:06 crc kubenswrapper[4808]: healthz check failed Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.486052 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7b9zg" podUID="9ebb9933-4652-4152-a529-5489fd9d5e5d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.539870 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:06 crc kubenswrapper[4808]: E1002 16:43:06.540745 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:07.040725253 +0000 UTC m=+154.366254253 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.542561 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-v4bld"] Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.551856 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ddlqv"] Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.567283 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmfsh"] Oct 02 16:43:06 crc kubenswrapper[4808]: W1002 16:43:06.592644 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c05ec42_ae20_46df_9231_a9a98a435103.slice/crio-4a36d18332b82a7219873bf74c22e584009481f9d9e979a3d7c6420560ce9585 WatchSource:0}: Error finding container 4a36d18332b82a7219873bf74c22e584009481f9d9e979a3d7c6420560ce9585: Status 404 returned error can't find the container with id 4a36d18332b82a7219873bf74c22e584009481f9d9e979a3d7c6420560ce9585 Oct 02 16:43:06 crc kubenswrapper[4808]: W1002 16:43:06.620182 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c98e3a0_cf25_4f21_acff_13035424b72f.slice/crio-761b3f3978a8899921c594439e2cac2fa9e41c4632e87b9059ab2fb98ecb2a79 WatchSource:0}: Error finding container 761b3f3978a8899921c594439e2cac2fa9e41c4632e87b9059ab2fb98ecb2a79: Status 404 returned error can't find the container with id 761b3f3978a8899921c594439e2cac2fa9e41c4632e87b9059ab2fb98ecb2a79 Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.627711 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zqnpj" event={"ID":"3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e","Type":"ContainerStarted","Data":"a9592964126e52acb3db7b640e1cf93853032bb70133e6dd7988c1b23f8a823f"} Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.627805 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zqnpj" event={"ID":"3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e","Type":"ContainerStarted","Data":"25e28e775dba047e5f5d252dae8517041c5c8c4b4fe784b4db3a2d36c6abd559"} Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.633565 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-lkxxm" event={"ID":"2ea7e7af-a0cd-4fda-8caf-b9db0c8e5205","Type":"ContainerStarted","Data":"e91e116dbfa4a4ddb27ec81625fad19a4051a0ae30ee1a91a09fafa9db7a4c97"} Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.645124 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bxvk8" event={"ID":"4d096969-852b-4401-bef0-757915f2cdb5","Type":"ContainerStarted","Data":"2d7f624ff0e00904714416f89d34543897c41e77ee5b63bfd68332e65424763a"} Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.653756 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:06 crc kubenswrapper[4808]: E1002 16:43:06.654106 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:07.154078132 +0000 UTC m=+154.479607132 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.679830 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ppxrn" event={"ID":"580bbc4a-b873-4631-86c3-476a771b9a70","Type":"ContainerStarted","Data":"40ecda264639a4d667f9a68aae477b73176be1759469b6b3142b6a8dbc7fb37f"} Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.682015 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ppxrn" Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.683422 4808 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-ppxrn container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.683689 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ppxrn" podUID="580bbc4a-b873-4631-86c3-476a771b9a70" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.705226 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kp2fm" event={"ID":"13d94271-250f-49f2-b1ef-4a97e991a124","Type":"ContainerStarted","Data":"0d31c5243c8744de415bbb6bb466f7ac777eb90004bf9c81479624ca2604e1db"} Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.716353 4808 generic.go:334] "Generic (PLEG): container finished" podID="67d706bc-9776-4edf-98e3-7eb03ff5ef75" containerID="d65df788d28664f889e98566717e14795bc00bec22d2bf075b934c10b210da9e" exitCode=0 Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.716519 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" event={"ID":"67d706bc-9776-4edf-98e3-7eb03ff5ef75","Type":"ContainerDied","Data":"d65df788d28664f889e98566717e14795bc00bec22d2bf075b934c10b210da9e"} Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.719292 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc" event={"ID":"05b1a2b9-4809-4e5b-87e3-49b61d3c5184","Type":"ContainerStarted","Data":"9d1a6cf64d30ea0a7f129c861b13fd9f8d3327ab87b94f2c9ebf4e9903d64550"} Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.720646 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmglv" event={"ID":"bf03d573-aaae-4b60-b285-da3bb7d3e4e3","Type":"ContainerStarted","Data":"fc388189131b1f8f4bfd7e939ff68f0f826621080bee0e96f51cadbaf38d8b67"} Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.730551 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8974h" event={"ID":"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324","Type":"ContainerStarted","Data":"509f66c4aee579b514a27a4db21816ff20c9e09aaf37c319cbb4a7addffac342"} Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.730597 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.737468 4808 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-8974h container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" start-of-body= Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.737549 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-8974h" podUID="8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.737809 4808 generic.go:334] "Generic (PLEG): container finished" podID="e615c64c-94fd-44fa-a119-69435c37bd82" containerID="63a12f27827fec2bdd44df3cba76d74027bb39aee01991f9771203ec826f007a" exitCode=0 Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.738078 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" event={"ID":"e615c64c-94fd-44fa-a119-69435c37bd82","Type":"ContainerDied","Data":"63a12f27827fec2bdd44df3cba76d74027bb39aee01991f9771203ec826f007a"} Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.742163 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-tj972" event={"ID":"7777a7df-de5d-4287-a6e9-bcb70a07fdfd","Type":"ContainerStarted","Data":"87a95e5be676efca7581cdd74d2b6cd408c347a0150dd78f544a13e25d2de74c"} Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.744555 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2htct" event={"ID":"6ae1cc03-9560-4e85-889e-63eda4518302","Type":"ContainerStarted","Data":"c87c00ad3c878ed5f781988cf0cf1cef67b2bd038a08640f64ddf6197d8de97a"} Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.744587 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2htct" event={"ID":"6ae1cc03-9560-4e85-889e-63eda4518302","Type":"ContainerStarted","Data":"2d091cd5371048bf961922aedb41a285064233c32b750180a608c2c8f67e5cfe"} Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.747105 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-7b9zg" event={"ID":"9ebb9933-4652-4152-a529-5489fd9d5e5d","Type":"ContainerStarted","Data":"58d6f1f20f435223ddec6b373c3ca1a05896f97a84a3d89c3336ef74cda12486"} Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.753707 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sksnr" event={"ID":"646a1a14-e3d6-43e8-92b8-3ae7d44e96da","Type":"ContainerStarted","Data":"e6ea559f2484b9aff0dd922a7bc9d577d3737a65aaf8e749739ffbe47a7e9c50"} Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.754834 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:06 crc kubenswrapper[4808]: E1002 16:43:06.759696 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:07.259680036 +0000 UTC m=+154.585209036 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.759980 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r4vs4"] Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.760645 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" event={"ID":"862243ee-b010-419a-884a-0824520ae00d","Type":"ContainerStarted","Data":"dacfc703a7e615cac96b4cc7b619998e61b36429a128a831bf2f1ca7b1942c91"} Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.778362 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-vh4gp"] Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.783032 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmglv" podStartSLOduration=132.783002932 podStartE2EDuration="2m12.783002932s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:06.777501554 +0000 UTC m=+154.103030554" watchObservedRunningTime="2025-10-02 16:43:06.783002932 +0000 UTC m=+154.108531932" Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.786394 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf" event={"ID":"f6a13086-47b7-4624-bd50-773a4badcb3f","Type":"ContainerStarted","Data":"1137feb48a70c1fe150df13db74ecff0d322778bc0d836d6e921af495294b1ee"} Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.786453 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf" event={"ID":"f6a13086-47b7-4624-bd50-773a4badcb3f","Type":"ContainerStarted","Data":"2114951eb9acad76a169a86c648bfa853f1b25104e523db7aeebaa7e892621d7"} Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.786472 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz"] Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.789251 4808 patch_prober.go:28] interesting pod/downloads-7954f5f757-qsqkp container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.789326 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qsqkp" podUID="8bc8bd65-c754-4a0a-9f11-40d444e2492a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.805838 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9"] Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.810826 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lbw7q" podStartSLOduration=132.81079385 podStartE2EDuration="2m12.81079385s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:06.806764619 +0000 UTC m=+154.132293639" watchObservedRunningTime="2025-10-02 16:43:06.81079385 +0000 UTC m=+154.136322850" Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.847162 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-7b9zg" podStartSLOduration=132.847140143 podStartE2EDuration="2m12.847140143s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:06.846692242 +0000 UTC m=+154.172221242" watchObservedRunningTime="2025-10-02 16:43:06.847140143 +0000 UTC m=+154.172669143" Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.858030 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:06 crc kubenswrapper[4808]: E1002 16:43:06.858473 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:07.358447458 +0000 UTC m=+154.683976458 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.860482 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:06 crc kubenswrapper[4808]: E1002 16:43:06.860819 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:07.360805797 +0000 UTC m=+154.686334857 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.874444 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw"] Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.882609 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-bxvk8" podStartSLOduration=4.882592504 podStartE2EDuration="4.882592504s" podCreationTimestamp="2025-10-02 16:43:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:06.881927548 +0000 UTC m=+154.207456548" watchObservedRunningTime="2025-10-02 16:43:06.882592504 +0000 UTC m=+154.208121504" Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.937854 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qc629"] Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.954162 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-ffv8w" podStartSLOduration=132.954129872 podStartE2EDuration="2m12.954129872s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:06.940753386 +0000 UTC m=+154.266282376" watchObservedRunningTime="2025-10-02 16:43:06.954129872 +0000 UTC m=+154.279658862" Oct 02 16:43:06 crc kubenswrapper[4808]: I1002 16:43:06.962928 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:06 crc kubenswrapper[4808]: E1002 16:43:06.968557 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:07.468531894 +0000 UTC m=+154.794060884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.019341 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2htks"] Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.035552 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-8974h" podStartSLOduration=133.035519697 podStartE2EDuration="2m13.035519697s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:06.985043699 +0000 UTC m=+154.310572699" watchObservedRunningTime="2025-10-02 16:43:07.035519697 +0000 UTC m=+154.361048707" Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.044181 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4nnw"] Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.065008 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-p47w4"] Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.065067 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq"] Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.065078 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-7pnwt"] Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.070578 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-4ff6b"] Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.076026 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.076213 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gpxn2"] Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.077130 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-q7ck8"] Oct 02 16:43:07 crc kubenswrapper[4808]: E1002 16:43:07.077542 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:07.577522113 +0000 UTC m=+154.903051113 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.078458 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-lkxxm" podStartSLOduration=135.078435546 podStartE2EDuration="2m15.078435546s" podCreationTimestamp="2025-10-02 16:40:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:07.033965878 +0000 UTC m=+154.359494878" watchObservedRunningTime="2025-10-02 16:43:07.078435546 +0000 UTC m=+154.403964536" Oct 02 16:43:07 crc kubenswrapper[4808]: W1002 16:43:07.110681 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e7f5669_a998_4474_a5cc_47a075e7658f.slice/crio-e2f9caaae31f07392ec97757d50c15bc6c24a5d9b0d46df036e32a7fef63d186 WatchSource:0}: Error finding container e2f9caaae31f07392ec97757d50c15bc6c24a5d9b0d46df036e32a7fef63d186: Status 404 returned error can't find the container with id e2f9caaae31f07392ec97757d50c15bc6c24a5d9b0d46df036e32a7fef63d186 Oct 02 16:43:07 crc kubenswrapper[4808]: W1002 16:43:07.111794 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod483da6c8_be0b_4025_9ccb_defe286e2d0e.slice/crio-c75a59760bbc1156ee686af14554f0349a54f8e6f7a3d7ca7a8f2f0c416c1deb WatchSource:0}: Error finding container c75a59760bbc1156ee686af14554f0349a54f8e6f7a3d7ca7a8f2f0c416c1deb: Status 404 returned error can't find the container with id c75a59760bbc1156ee686af14554f0349a54f8e6f7a3d7ca7a8f2f0c416c1deb Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.154202 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ppxrn" podStartSLOduration=133.13711256 podStartE2EDuration="2m13.13711256s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:07.1354986 +0000 UTC m=+154.461027620" watchObservedRunningTime="2025-10-02 16:43:07.13711256 +0000 UTC m=+154.462641560" Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.155039 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-dlddx" podStartSLOduration=133.15502629 podStartE2EDuration="2m13.15502629s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:07.104216423 +0000 UTC m=+154.429745433" watchObservedRunningTime="2025-10-02 16:43:07.15502629 +0000 UTC m=+154.480555300" Oct 02 16:43:07 crc kubenswrapper[4808]: W1002 16:43:07.165721 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee194b74_bef2_4385_8135_7c25d5de5589.slice/crio-aa59f360bae932660b217c8d41cf2e43607ab77216d35b06eb274c312ba261ae WatchSource:0}: Error finding container aa59f360bae932660b217c8d41cf2e43607ab77216d35b06eb274c312ba261ae: Status 404 returned error can't find the container with id aa59f360bae932660b217c8d41cf2e43607ab77216d35b06eb274c312ba261ae Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.169439 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4dk9g" podStartSLOduration=133.169420362 podStartE2EDuration="2m13.169420362s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:07.158725803 +0000 UTC m=+154.484254803" watchObservedRunningTime="2025-10-02 16:43:07.169420362 +0000 UTC m=+154.494949362" Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.178978 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:07 crc kubenswrapper[4808]: E1002 16:43:07.179456 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:07.679434554 +0000 UTC m=+155.004963564 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.227146 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2htct" podStartSLOduration=133.227126882 podStartE2EDuration="2m13.227126882s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:07.225431629 +0000 UTC m=+154.550960629" watchObservedRunningTime="2025-10-02 16:43:07.227126882 +0000 UTC m=+154.552655882" Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.281571 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:07 crc kubenswrapper[4808]: E1002 16:43:07.282381 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:07.78235596 +0000 UTC m=+155.107884960 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.393041 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:07 crc kubenswrapper[4808]: E1002 16:43:07.393572 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:07.893550984 +0000 UTC m=+155.219079984 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.481930 4808 patch_prober.go:28] interesting pod/router-default-5444994796-7b9zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 16:43:07 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Oct 02 16:43:07 crc kubenswrapper[4808]: [+]process-running ok Oct 02 16:43:07 crc kubenswrapper[4808]: healthz check failed Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.482008 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7b9zg" podUID="9ebb9933-4652-4152-a529-5489fd9d5e5d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.500096 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:07 crc kubenswrapper[4808]: E1002 16:43:07.500855 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:08.0008264 +0000 UTC m=+155.326355570 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.601295 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:07 crc kubenswrapper[4808]: E1002 16:43:07.601488 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:08.101449748 +0000 UTC m=+155.426978748 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.602560 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:07 crc kubenswrapper[4808]: E1002 16:43:07.602999 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:08.102976077 +0000 UTC m=+155.428505077 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.703990 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:07 crc kubenswrapper[4808]: E1002 16:43:07.704455 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:08.204438186 +0000 UTC m=+155.529967186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.807574 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:07 crc kubenswrapper[4808]: E1002 16:43:07.808055 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:08.30803824 +0000 UTC m=+155.633567240 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.816361 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ddlqv" event={"ID":"3c05ec42-ae20-46df-9231-a9a98a435103","Type":"ContainerStarted","Data":"5224dad47e8fd558090199a6160e7e4dd05718eea4c5e71e24cfae944be475f6"} Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.817120 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ddlqv" event={"ID":"3c05ec42-ae20-46df-9231-a9a98a435103","Type":"ContainerStarted","Data":"4a36d18332b82a7219873bf74c22e584009481f9d9e979a3d7c6420560ce9585"} Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.825418 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r4vs4" event={"ID":"4d89d791-944c-4afc-b0f9-83d3f7604132","Type":"ContainerStarted","Data":"f9ac810ec805f5eaca743e9c388178bb908587f84ac699aacd49b288866516a0"} Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.831219 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw" event={"ID":"af55464e-2244-46e0-a2bc-9211829b2d7d","Type":"ContainerStarted","Data":"9ed3e509e7b8032cf79b496b373898d86249af6a29804e8770ddf820d3277589"} Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.836930 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" event={"ID":"8e7f5669-a998-4474-a5cc-47a075e7658f","Type":"ContainerStarted","Data":"e2f9caaae31f07392ec97757d50c15bc6c24a5d9b0d46df036e32a7fef63d186"} Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.849894 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zqnpj" event={"ID":"3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e","Type":"ContainerStarted","Data":"938e14ae3329c72085279b594339ad65b71598590e633582f1ab9adf7e3e28e6"} Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.863782 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4bld" event={"ID":"4c98e3a0-cf25-4f21-acff-13035424b72f","Type":"ContainerStarted","Data":"348d228055ab1e5c819578c8d95e890a0d7ec8516ec7fc3e1d77ffccb7e6af6b"} Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.863838 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4bld" event={"ID":"4c98e3a0-cf25-4f21-acff-13035424b72f","Type":"ContainerStarted","Data":"761b3f3978a8899921c594439e2cac2fa9e41c4632e87b9059ab2fb98ecb2a79"} Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.879851 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-zqnpj" podStartSLOduration=133.879827753 podStartE2EDuration="2m13.879827753s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:07.872484109 +0000 UTC m=+155.198013109" watchObservedRunningTime="2025-10-02 16:43:07.879827753 +0000 UTC m=+155.205356754" Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.884762 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kp2fm" event={"ID":"13d94271-250f-49f2-b1ef-4a97e991a124","Type":"ContainerStarted","Data":"fa7ed7ed73c4e38ffeb6fb2292f095328cff58cdf9c0ed92fa0ed0134391189b"} Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.889170 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2htks" event={"ID":"681e26e4-4764-43f4-b3f9-d084f9619e5f","Type":"ContainerStarted","Data":"b483d87608e405d2b583f014b9d8cbd41f0961e994f0f5896e054e87a677239b"} Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.904693 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4bld" podStartSLOduration=132.904660258 podStartE2EDuration="2m12.904660258s" podCreationTimestamp="2025-10-02 16:40:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:07.894644696 +0000 UTC m=+155.220173716" watchObservedRunningTime="2025-10-02 16:43:07.904660258 +0000 UTC m=+155.230189278" Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.908612 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:07 crc kubenswrapper[4808]: E1002 16:43:07.908795 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:08.4087562 +0000 UTC m=+155.734285200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.909014 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:07 crc kubenswrapper[4808]: E1002 16:43:07.911048 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:08.411025317 +0000 UTC m=+155.736554317 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.913576 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-7pnwt" event={"ID":"b3771709-b077-4820-bfb3-1ea9986866bb","Type":"ContainerStarted","Data":"31cbabff6c4a6902c71ab890b720653e208f071ba589fe6239d7aebc52e26205"} Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.919159 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kp2fm" podStartSLOduration=133.919137391 podStartE2EDuration="2m13.919137391s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:07.918554707 +0000 UTC m=+155.244083707" watchObservedRunningTime="2025-10-02 16:43:07.919137391 +0000 UTC m=+155.244666391" Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.954809 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4nnw" event={"ID":"483da6c8-be0b-4025-9ccb-defe286e2d0e","Type":"ContainerStarted","Data":"c75a59760bbc1156ee686af14554f0349a54f8e6f7a3d7ca7a8f2f0c416c1deb"} Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.979974 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" event={"ID":"02097ffc-aa46-438f-b11d-c98301f3e617","Type":"ContainerStarted","Data":"fb52e103568780b100651c7ec7e80b7c761d63ba21dc16dd926c6b13d584b060"} Oct 02 16:43:07 crc kubenswrapper[4808]: I1002 16:43:07.995227 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-4ff6b" event={"ID":"ee194b74-bef2-4385-8135-7c25d5de5589","Type":"ContainerStarted","Data":"aa59f360bae932660b217c8d41cf2e43607ab77216d35b06eb274c312ba261ae"} Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.007064 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc" event={"ID":"05b1a2b9-4809-4e5b-87e3-49b61d3c5184","Type":"ContainerStarted","Data":"bb64782c75f2f750c7e4de370e57811ba7d2fa100b9febfca723c15463650eae"} Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.009862 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:08 crc kubenswrapper[4808]: E1002 16:43:08.010262 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:08.5102268 +0000 UTC m=+155.835755800 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.027685 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz" event={"ID":"aea5ac39-3315-4186-9002-e836172b6d34","Type":"ContainerStarted","Data":"787f43c20f31552485ae15d355761312011748920feced49ab2be50f50268f3e"} Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.032535 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz" event={"ID":"aea5ac39-3315-4186-9002-e836172b6d34","Type":"ContainerStarted","Data":"f5327c49d869e2c1d38341a336c7781e8d679bffd694affa4e586bc8b8fc2469"} Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.043769 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p47w4" event={"ID":"9e67480b-9e3f-433f-95f1-ce1e32298fdc","Type":"ContainerStarted","Data":"61e066558e004c24a69cfa2d9e83569a3039326abff5f0ae4252d045ab9b23b4"} Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.055261 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" event={"ID":"862243ee-b010-419a-884a-0824520ae00d","Type":"ContainerStarted","Data":"b4ce5d6b3bca25999f28bf36169f5e81eb90dc05cf444ae9b91c0caf21df1c0c"} Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.056163 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.062518 4808 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-6br9m container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.062577 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" podUID="862243ee-b010-419a-884a-0824520ae00d" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.086574 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" podStartSLOduration=133.086551288 podStartE2EDuration="2m13.086551288s" podCreationTimestamp="2025-10-02 16:40:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:08.086007234 +0000 UTC m=+155.411536234" watchObservedRunningTime="2025-10-02 16:43:08.086551288 +0000 UTC m=+155.412080288" Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.088449 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sksnr" event={"ID":"646a1a14-e3d6-43e8-92b8-3ae7d44e96da","Type":"ContainerStarted","Data":"0baa21287766df9eb34a5bd84f1d6428cb50bae582f7b272a5e7baf0b0715c20"} Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.091599 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" event={"ID":"c74723a8-81cf-4633-8655-e1460025f1b7","Type":"ContainerStarted","Data":"4f4a104ef7859ab84dd7653986c84723fe3a1d0adcde8aaced7514f5f1d59a92"} Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.102520 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf" event={"ID":"f6a13086-47b7-4624-bd50-773a4badcb3f","Type":"ContainerStarted","Data":"d0718a80540541503226a69cf283407c45ab00304c1df6f385d4a2acd3f2117e"} Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.112204 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:08 crc kubenswrapper[4808]: E1002 16:43:08.114272 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:08.614256164 +0000 UTC m=+155.939785164 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.115290 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vh4gp" event={"ID":"7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1","Type":"ContainerStarted","Data":"7305aa29dd1b599d1e055f69d0df5e20e6c4788071c3f399821cab97a05a41c7"} Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.121528 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sksnr" podStartSLOduration=134.121494316 podStartE2EDuration="2m14.121494316s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:08.116685235 +0000 UTC m=+155.442214235" watchObservedRunningTime="2025-10-02 16:43:08.121494316 +0000 UTC m=+155.447023316" Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.133817 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmfsh" event={"ID":"2849754c-63a9-481f-8c8b-f328a0daaa2c","Type":"ContainerStarted","Data":"d169ff69cc6bdc3a017d0b81e3eb8cd388b24edb27c6f8ea378630848d4662b0"} Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.133871 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmfsh" event={"ID":"2849754c-63a9-481f-8c8b-f328a0daaa2c","Type":"ContainerStarted","Data":"e48bcc083a52c097efcc0710d9c66b8500f7a2ec992d1678e70d2e6048dd5077"} Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.138467 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9" event={"ID":"e736d6e6-545b-451a-80d8-9f4e6d83b4bc","Type":"ContainerStarted","Data":"fab7cf9f3dfa89ed02ce8003aa158d8d9098e78c8580426c6841bc481c6bd2e8"} Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.138531 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9" event={"ID":"e736d6e6-545b-451a-80d8-9f4e6d83b4bc","Type":"ContainerStarted","Data":"c92e19798643f1ebdf444c7f3cde8ee2d1a9295fe758d7e15d7c213ea3504a0a"} Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.139562 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9" Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.141398 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6vlpf" podStartSLOduration=134.141376006 podStartE2EDuration="2m14.141376006s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:08.137536379 +0000 UTC m=+155.463065409" watchObservedRunningTime="2025-10-02 16:43:08.141376006 +0000 UTC m=+155.466905006" Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.154417 4808 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6dtm9 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.154496 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9" podUID="e736d6e6-545b-451a-80d8-9f4e6d83b4bc" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.193257 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" event={"ID":"8ddfee45-b557-4f8b-8702-6896ca1fa9ff","Type":"ContainerStarted","Data":"121828a74b36baee2b798f8f2c0c85811b5f55cdf19745e5b3b75ed730c5a272"} Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.194478 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.199607 4808 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-dp2fq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:5443/healthz\": dial tcp 10.217.0.42:5443: connect: connection refused" start-of-body= Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.199682 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" podUID="8ddfee45-b557-4f8b-8702-6896ca1fa9ff" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.42:5443/healthz\": dial tcp 10.217.0.42:5443: connect: connection refused" Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.219871 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.221905 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-tj972" event={"ID":"7777a7df-de5d-4287-a6e9-bcb70a07fdfd","Type":"ContainerStarted","Data":"05047216b64a11085572105b6c55c464c056c4ae08f87bf0b7436516e241e0e8"} Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.223705 4808 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-8974h container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" start-of-body= Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.223905 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-8974h" podUID="8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" Oct 02 16:43:08 crc kubenswrapper[4808]: E1002 16:43:08.227254 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:08.727206832 +0000 UTC m=+156.052735832 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.245552 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmfsh" podStartSLOduration=134.245528563 podStartE2EDuration="2m14.245528563s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:08.164087156 +0000 UTC m=+155.489616156" watchObservedRunningTime="2025-10-02 16:43:08.245528563 +0000 UTC m=+155.571057593" Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.247796 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9" podStartSLOduration=134.247780179 podStartE2EDuration="2m14.247780179s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:08.218410041 +0000 UTC m=+155.543939041" watchObservedRunningTime="2025-10-02 16:43:08.247780179 +0000 UTC m=+155.573309179" Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.287940 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ppxrn" Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.324587 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:08 crc kubenswrapper[4808]: E1002 16:43:08.330086 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:08.830065266 +0000 UTC m=+156.155594266 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.339645 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" podStartSLOduration=134.339622036 podStartE2EDuration="2m14.339622036s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:08.308830634 +0000 UTC m=+155.634359634" watchObservedRunningTime="2025-10-02 16:43:08.339622036 +0000 UTC m=+155.665151036" Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.428435 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:08 crc kubenswrapper[4808]: E1002 16:43:08.429142 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:08.929120465 +0000 UTC m=+156.254649465 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.481293 4808 patch_prober.go:28] interesting pod/router-default-5444994796-7b9zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 16:43:08 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Oct 02 16:43:08 crc kubenswrapper[4808]: [+]process-running ok Oct 02 16:43:08 crc kubenswrapper[4808]: healthz check failed Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.481359 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7b9zg" podUID="9ebb9933-4652-4152-a529-5489fd9d5e5d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.530436 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:08 crc kubenswrapper[4808]: E1002 16:43:08.530832 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:09.030817081 +0000 UTC m=+156.356346081 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.633843 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:08 crc kubenswrapper[4808]: E1002 16:43:08.634116 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:09.134081146 +0000 UTC m=+156.459610146 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.634580 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:08 crc kubenswrapper[4808]: E1002 16:43:08.635059 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:09.13503919 +0000 UTC m=+156.460568190 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.735444 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:08 crc kubenswrapper[4808]: E1002 16:43:08.735617 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:09.235588036 +0000 UTC m=+156.561117046 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.735697 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:08 crc kubenswrapper[4808]: E1002 16:43:08.736042 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:09.236033447 +0000 UTC m=+156.561562447 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.836469 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:08 crc kubenswrapper[4808]: E1002 16:43:08.836714 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:09.336665246 +0000 UTC m=+156.662194256 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.836820 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:08 crc kubenswrapper[4808]: E1002 16:43:08.837177 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:09.337168469 +0000 UTC m=+156.662697469 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.938360 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:08 crc kubenswrapper[4808]: E1002 16:43:08.938571 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:09.438535176 +0000 UTC m=+156.764064176 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:08 crc kubenswrapper[4808]: I1002 16:43:08.939119 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:08 crc kubenswrapper[4808]: E1002 16:43:08.939534 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:09.439525311 +0000 UTC m=+156.765054311 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.040622 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:09 crc kubenswrapper[4808]: E1002 16:43:09.041455 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:09.541417611 +0000 UTC m=+156.866946621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.143251 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:09 crc kubenswrapper[4808]: E1002 16:43:09.143657 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:09.64364077 +0000 UTC m=+156.969169770 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.168152 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lbw7q" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.230654 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" event={"ID":"67d706bc-9776-4edf-98e3-7eb03ff5ef75","Type":"ContainerStarted","Data":"8998f55d717867233c5bd7e29db4d1a864398933096fe2fb60a6c9ec271cbcb2"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.231085 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" event={"ID":"67d706bc-9776-4edf-98e3-7eb03ff5ef75","Type":"ContainerStarted","Data":"bed76c715c55f16d03beafb50b9f68ec54b9d6242212a2974aed5befef453744"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.232604 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p47w4" event={"ID":"9e67480b-9e3f-433f-95f1-ce1e32298fdc","Type":"ContainerStarted","Data":"24c69f22a84d50da471c26e886e30dcbfdd7ec153719557e62a08135009c5458"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.232657 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p47w4" event={"ID":"9e67480b-9e3f-433f-95f1-ce1e32298fdc","Type":"ContainerStarted","Data":"a6f732f62cd0adc7a04580582ee069ea3dd4278f7788ebe8b5cf9a25fdf075ed"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.233863 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4nnw" event={"ID":"483da6c8-be0b-4025-9ccb-defe286e2d0e","Type":"ContainerStarted","Data":"5d6a1f0a08ee5ea6bfbfb581678c80c071f632d89d73ea4e4fc1d71de276194f"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.235471 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vh4gp" event={"ID":"7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1","Type":"ContainerStarted","Data":"a95c17cdacef5cad075f523d4c816c320392c3cb0fcc73e0e4737025ae2af9bb"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.235499 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vh4gp" event={"ID":"7f5e2e4e-cca0-4e1c-875b-f8da5085b7c1","Type":"ContainerStarted","Data":"d3a807d20f3111dd8c138fc911104efacee84987406dd2f728adb465465f9e04"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.235881 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-vh4gp" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.237171 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" event={"ID":"c74723a8-81cf-4633-8655-e1460025f1b7","Type":"ContainerStarted","Data":"8922f4f2f17ad396033076ece76e9ee8982c4f57182c3896a70dca03d9c2442a"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.237841 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.238961 4808 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-qc629 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.239006 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" podUID="c74723a8-81cf-4633-8655-e1460025f1b7" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.240949 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" event={"ID":"e615c64c-94fd-44fa-a119-69435c37bd82","Type":"ContainerStarted","Data":"d50e44edd05f9a3d9979bd7aa7eeeadce5ea5eafb8d549a4a864dce33030904d"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.242897 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r4vs4" event={"ID":"4d89d791-944c-4afc-b0f9-83d3f7604132","Type":"ContainerStarted","Data":"a0bce43ae5e9595d557e6810c42316d47fe61a4ab66f90f6042620fe3a2547e0"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.242940 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r4vs4" event={"ID":"4d89d791-944c-4afc-b0f9-83d3f7604132","Type":"ContainerStarted","Data":"4f94e12f5a9fbb2f57191e58f1a4378f8a189e6c7ac68ca7b685f16d4b88f48a"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.243408 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r4vs4" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.244484 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:09 crc kubenswrapper[4808]: E1002 16:43:09.244653 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:09.744625458 +0000 UTC m=+157.070154458 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.244735 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:09 crc kubenswrapper[4808]: E1002 16:43:09.245066 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:09.745058889 +0000 UTC m=+157.070588089 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.245154 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw" event={"ID":"af55464e-2244-46e0-a2bc-9211829b2d7d","Type":"ContainerStarted","Data":"b7bc0a3b03df07ebd5ed98d2c6a60e3dbbebd024ce364c504ef9303785fde32f"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.247889 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-7pnwt" event={"ID":"b3771709-b077-4820-bfb3-1ea9986866bb","Type":"ContainerStarted","Data":"6a24f90d270d75e24c9166b89633dca5ed7b1d34db8f02c16eeaf80d613c83c0"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.250008 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" event={"ID":"02097ffc-aa46-438f-b11d-c98301f3e617","Type":"ContainerStarted","Data":"c4917adf048a4ec961ca2f6c21b80a0dddab2f112a5bc63ba2a21834b63ce6cc"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.250774 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.252410 4808 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gpxn2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.252448 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" podUID="02097ffc-aa46-438f-b11d-c98301f3e617" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.253629 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ddlqv" event={"ID":"3c05ec42-ae20-46df-9231-a9a98a435103","Type":"ContainerStarted","Data":"1962527993c9658ab53b063ae91281cc5fb8fc97271972eb6062cf3a03a2f8f7"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.255587 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-tj972" event={"ID":"7777a7df-de5d-4287-a6e9-bcb70a07fdfd","Type":"ContainerStarted","Data":"8353a61cc992f0164f4e2e8624afaf740ff733045396361f3377bbb9a7c1337c"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.258945 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" event={"ID":"8e7f5669-a998-4474-a5cc-47a075e7658f","Type":"ContainerStarted","Data":"34626b6ff838c6ceda18fdbfd63c8b7e8e007edc2450259387ab8c72ee5cb100"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.261746 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" event={"ID":"8ddfee45-b557-4f8b-8702-6896ca1fa9ff","Type":"ContainerStarted","Data":"ab8fa85fde9abd635a2071d69db097ea02c69207d25b2d26de601f5320624708"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.262025 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" podStartSLOduration=135.262000254 podStartE2EDuration="2m15.262000254s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:09.260310502 +0000 UTC m=+156.585839502" watchObservedRunningTime="2025-10-02 16:43:09.262000254 +0000 UTC m=+156.587529254" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.262715 4808 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-dp2fq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:5443/healthz\": dial tcp 10.217.0.42:5443: connect: connection refused" start-of-body= Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.262770 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" podUID="8ddfee45-b557-4f8b-8702-6896ca1fa9ff" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.42:5443/healthz\": dial tcp 10.217.0.42:5443: connect: connection refused" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.265363 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-4ff6b" event={"ID":"ee194b74-bef2-4385-8135-7c25d5de5589","Type":"ContainerStarted","Data":"73fd4aa65d6f42c763951c238f73b94dafca219d89ec2ebd0fe6085f4b4a38e3"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.267715 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc" event={"ID":"05b1a2b9-4809-4e5b-87e3-49b61d3c5184","Type":"ContainerStarted","Data":"1127d7654166cf88c6a302686d29623d955f38df89233f091ee95593c32bd843"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.270415 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.270487 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.270929 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz" event={"ID":"aea5ac39-3315-4186-9002-e836172b6d34","Type":"ContainerStarted","Data":"9274f2c6cedab4be0425ef30921cd7a6485d894aa3ded81941e84bcc03c0bb04"} Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.271708 4808 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6dtm9 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.271761 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9" podUID="e736d6e6-545b-451a-80d8-9f4e6d83b4bc" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.271882 4808 patch_prober.go:28] interesting pod/apiserver-76f77b778f-2bgkb container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.10:8443/livez\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.271942 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" podUID="67d706bc-9776-4edf-98e3-7eb03ff5ef75" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.10:8443/livez\": dial tcp 10.217.0.10:8443: connect: connection refused" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.289730 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r4vs4" podStartSLOduration=135.289708641 podStartE2EDuration="2m15.289708641s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:09.287242389 +0000 UTC m=+156.612771389" watchObservedRunningTime="2025-10-02 16:43:09.289708641 +0000 UTC m=+156.615237641" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.309339 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-q7ck8" podStartSLOduration=135.309316583 podStartE2EDuration="2m15.309316583s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:09.308946044 +0000 UTC m=+156.634475064" watchObservedRunningTime="2025-10-02 16:43:09.309316583 +0000 UTC m=+156.634845583" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.325030 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-ddlqv" podStartSLOduration=135.325009288 podStartE2EDuration="2m15.325009288s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:09.325008758 +0000 UTC m=+156.650537778" watchObservedRunningTime="2025-10-02 16:43:09.325009288 +0000 UTC m=+156.650538288" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.347493 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.347979 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.348457 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.349550 4808 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-5wvmf container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.30:8443/livez\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.349618 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" podUID="e615c64c-94fd-44fa-a119-69435c37bd82" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.30:8443/livez\": dial tcp 10.217.0.30:8443: connect: connection refused" Oct 02 16:43:09 crc kubenswrapper[4808]: E1002 16:43:09.350878 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:09.850853157 +0000 UTC m=+157.176382277 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.383101 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw" podStartSLOduration=136.383079797 podStartE2EDuration="2m16.383079797s" podCreationTimestamp="2025-10-02 16:40:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:09.347298868 +0000 UTC m=+156.672827868" watchObservedRunningTime="2025-10-02 16:43:09.383079797 +0000 UTC m=+156.708608797" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.383889 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p47w4" podStartSLOduration=135.383885817 podStartE2EDuration="2m15.383885817s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:09.38199718 +0000 UTC m=+156.707526180" watchObservedRunningTime="2025-10-02 16:43:09.383885817 +0000 UTC m=+156.709414807" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.401172 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-vh4gp" podStartSLOduration=7.401154451 podStartE2EDuration="7.401154451s" podCreationTimestamp="2025-10-02 16:43:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:09.400248838 +0000 UTC m=+156.725777838" watchObservedRunningTime="2025-10-02 16:43:09.401154451 +0000 UTC m=+156.726683451" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.436408 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" podStartSLOduration=135.436379016 podStartE2EDuration="2m15.436379016s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:09.432266143 +0000 UTC m=+156.757795153" watchObservedRunningTime="2025-10-02 16:43:09.436379016 +0000 UTC m=+156.761908016" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.451657 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:09 crc kubenswrapper[4808]: E1002 16:43:09.452180 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:09.952160793 +0000 UTC m=+157.277689793 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.457940 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4nnw" podStartSLOduration=135.457918418 podStartE2EDuration="2m15.457918418s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:09.455303952 +0000 UTC m=+156.780832952" watchObservedRunningTime="2025-10-02 16:43:09.457918418 +0000 UTC m=+156.783447418" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.480541 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" podStartSLOduration=135.480522516 podStartE2EDuration="2m15.480522516s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:09.478560586 +0000 UTC m=+156.804089586" watchObservedRunningTime="2025-10-02 16:43:09.480522516 +0000 UTC m=+156.806051516" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.483573 4808 patch_prober.go:28] interesting pod/router-default-5444994796-7b9zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 16:43:09 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Oct 02 16:43:09 crc kubenswrapper[4808]: [+]process-running ok Oct 02 16:43:09 crc kubenswrapper[4808]: healthz check failed Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.483640 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7b9zg" podUID="9ebb9933-4652-4152-a529-5489fd9d5e5d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.499202 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-tj972" podStartSLOduration=135.499180154 podStartE2EDuration="2m15.499180154s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:09.497304747 +0000 UTC m=+156.822833757" watchObservedRunningTime="2025-10-02 16:43:09.499180154 +0000 UTC m=+156.824709154" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.553117 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:09 crc kubenswrapper[4808]: E1002 16:43:09.553602 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:10.053573891 +0000 UTC m=+157.379102891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.568959 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" podStartSLOduration=135.568936327 podStartE2EDuration="2m15.568936327s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:09.533571999 +0000 UTC m=+156.859100999" watchObservedRunningTime="2025-10-02 16:43:09.568936327 +0000 UTC m=+156.894465327" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.614011 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-4ff6b" podStartSLOduration=8.613990249 podStartE2EDuration="8.613990249s" podCreationTimestamp="2025-10-02 16:43:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:09.610009559 +0000 UTC m=+156.935538559" watchObservedRunningTime="2025-10-02 16:43:09.613990249 +0000 UTC m=+156.939519249" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.614577 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-7pnwt" podStartSLOduration=134.614572014 podStartE2EDuration="2m14.614572014s" podCreationTimestamp="2025-10-02 16:40:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:09.569734297 +0000 UTC m=+156.895263297" watchObservedRunningTime="2025-10-02 16:43:09.614572014 +0000 UTC m=+156.940101014" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.648468 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.654179 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:09 crc kubenswrapper[4808]: E1002 16:43:09.654767 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:10.154741193 +0000 UTC m=+157.480270353 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.692988 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s6jhz" podStartSLOduration=135.692963154 podStartE2EDuration="2m15.692963154s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:09.662885928 +0000 UTC m=+156.988414928" watchObservedRunningTime="2025-10-02 16:43:09.692963154 +0000 UTC m=+157.018492154" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.693213 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zxfsc" podStartSLOduration=135.69320608 podStartE2EDuration="2m15.69320608s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:09.690720038 +0000 UTC m=+157.016249038" watchObservedRunningTime="2025-10-02 16:43:09.69320608 +0000 UTC m=+157.018735080" Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.760852 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:09 crc kubenswrapper[4808]: E1002 16:43:09.761327 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:10.261300251 +0000 UTC m=+157.586829281 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.862770 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:09 crc kubenswrapper[4808]: E1002 16:43:09.863261 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:10.363217822 +0000 UTC m=+157.688746822 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.964115 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:09 crc kubenswrapper[4808]: E1002 16:43:09.964346 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:10.464309153 +0000 UTC m=+157.789838153 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:09 crc kubenswrapper[4808]: I1002 16:43:09.964903 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:09 crc kubenswrapper[4808]: E1002 16:43:09.965333 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:10.465317108 +0000 UTC m=+157.790846108 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.027111 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nzsf5"] Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.028665 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nzsf5" Oct 02 16:43:10 crc kubenswrapper[4808]: W1002 16:43:10.050249 4808 reflector.go:561] object-"openshift-marketplace"/"community-operators-dockercfg-dmngl": failed to list *v1.Secret: secrets "community-operators-dockercfg-dmngl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Oct 02 16:43:10 crc kubenswrapper[4808]: E1002 16:43:10.050306 4808 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"community-operators-dockercfg-dmngl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"community-operators-dockercfg-dmngl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.066465 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:10 crc kubenswrapper[4808]: E1002 16:43:10.066612 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:10.566583122 +0000 UTC m=+157.892112122 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.067010 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.067108 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c0f4487-6602-4c9d-b414-229588b2e797-utilities\") pod \"community-operators-nzsf5\" (UID: \"6c0f4487-6602-4c9d-b414-229588b2e797\") " pod="openshift-marketplace/community-operators-nzsf5" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.067127 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c0f4487-6602-4c9d-b414-229588b2e797-catalog-content\") pod \"community-operators-nzsf5\" (UID: \"6c0f4487-6602-4c9d-b414-229588b2e797\") " pod="openshift-marketplace/community-operators-nzsf5" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.067150 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pn24\" (UniqueName: \"kubernetes.io/projected/6c0f4487-6602-4c9d-b414-229588b2e797-kube-api-access-4pn24\") pod \"community-operators-nzsf5\" (UID: \"6c0f4487-6602-4c9d-b414-229588b2e797\") " pod="openshift-marketplace/community-operators-nzsf5" Oct 02 16:43:10 crc kubenswrapper[4808]: E1002 16:43:10.067598 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:10.567586358 +0000 UTC m=+157.893115358 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.079445 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nzsf5"] Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.168397 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:10 crc kubenswrapper[4808]: E1002 16:43:10.168598 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:10.668563885 +0000 UTC m=+157.994092885 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.168723 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.168817 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c0f4487-6602-4c9d-b414-229588b2e797-utilities\") pod \"community-operators-nzsf5\" (UID: \"6c0f4487-6602-4c9d-b414-229588b2e797\") " pod="openshift-marketplace/community-operators-nzsf5" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.168840 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c0f4487-6602-4c9d-b414-229588b2e797-catalog-content\") pod \"community-operators-nzsf5\" (UID: \"6c0f4487-6602-4c9d-b414-229588b2e797\") " pod="openshift-marketplace/community-operators-nzsf5" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.168866 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pn24\" (UniqueName: \"kubernetes.io/projected/6c0f4487-6602-4c9d-b414-229588b2e797-kube-api-access-4pn24\") pod \"community-operators-nzsf5\" (UID: \"6c0f4487-6602-4c9d-b414-229588b2e797\") " pod="openshift-marketplace/community-operators-nzsf5" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.169473 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c0f4487-6602-4c9d-b414-229588b2e797-utilities\") pod \"community-operators-nzsf5\" (UID: \"6c0f4487-6602-4c9d-b414-229588b2e797\") " pod="openshift-marketplace/community-operators-nzsf5" Oct 02 16:43:10 crc kubenswrapper[4808]: E1002 16:43:10.169490 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:10.669467408 +0000 UTC m=+157.994996408 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.169650 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c0f4487-6602-4c9d-b414-229588b2e797-catalog-content\") pod \"community-operators-nzsf5\" (UID: \"6c0f4487-6602-4c9d-b414-229588b2e797\") " pod="openshift-marketplace/community-operators-nzsf5" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.210349 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pn24\" (UniqueName: \"kubernetes.io/projected/6c0f4487-6602-4c9d-b414-229588b2e797-kube-api-access-4pn24\") pod \"community-operators-nzsf5\" (UID: \"6c0f4487-6602-4c9d-b414-229588b2e797\") " pod="openshift-marketplace/community-operators-nzsf5" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.252059 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zbj7g"] Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.253429 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zbj7g" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.263715 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.269874 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:10 crc kubenswrapper[4808]: E1002 16:43:10.271722 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:10.771700867 +0000 UTC m=+158.097229867 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.278139 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zbj7g"] Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.287540 4808 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gpxn2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.287599 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" podUID="02097ffc-aa46-438f-b11d-c98301f3e617" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.291402 4808 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-dp2fq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:5443/healthz\": dial tcp 10.217.0.42:5443: connect: connection refused" start-of-body= Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.292340 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" podUID="8ddfee45-b557-4f8b-8702-6896ca1fa9ff" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.42:5443/healthz\": dial tcp 10.217.0.42:5443: connect: connection refused" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.291425 4808 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-qc629 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.292543 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" podUID="c74723a8-81cf-4633-8655-e1460025f1b7" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.355150 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6dtm9" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.374629 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09726ef2-b31b-48cd-8577-7eb3ae6aaca6-utilities\") pod \"certified-operators-zbj7g\" (UID: \"09726ef2-b31b-48cd-8577-7eb3ae6aaca6\") " pod="openshift-marketplace/certified-operators-zbj7g" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.374680 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09726ef2-b31b-48cd-8577-7eb3ae6aaca6-catalog-content\") pod \"certified-operators-zbj7g\" (UID: \"09726ef2-b31b-48cd-8577-7eb3ae6aaca6\") " pod="openshift-marketplace/certified-operators-zbj7g" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.375006 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.375063 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tqpg\" (UniqueName: \"kubernetes.io/projected/09726ef2-b31b-48cd-8577-7eb3ae6aaca6-kube-api-access-6tqpg\") pod \"certified-operators-zbj7g\" (UID: \"09726ef2-b31b-48cd-8577-7eb3ae6aaca6\") " pod="openshift-marketplace/certified-operators-zbj7g" Oct 02 16:43:10 crc kubenswrapper[4808]: E1002 16:43:10.385861 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:10.885841205 +0000 UTC m=+158.211370205 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.427125 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lhm6x"] Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.429612 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lhm6x" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.478778 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.479462 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09726ef2-b31b-48cd-8577-7eb3ae6aaca6-utilities\") pod \"certified-operators-zbj7g\" (UID: \"09726ef2-b31b-48cd-8577-7eb3ae6aaca6\") " pod="openshift-marketplace/certified-operators-zbj7g" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.479506 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09726ef2-b31b-48cd-8577-7eb3ae6aaca6-catalog-content\") pod \"certified-operators-zbj7g\" (UID: \"09726ef2-b31b-48cd-8577-7eb3ae6aaca6\") " pod="openshift-marketplace/certified-operators-zbj7g" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.479672 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tqpg\" (UniqueName: \"kubernetes.io/projected/09726ef2-b31b-48cd-8577-7eb3ae6aaca6-kube-api-access-6tqpg\") pod \"certified-operators-zbj7g\" (UID: \"09726ef2-b31b-48cd-8577-7eb3ae6aaca6\") " pod="openshift-marketplace/certified-operators-zbj7g" Oct 02 16:43:10 crc kubenswrapper[4808]: E1002 16:43:10.480200 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:10.980164005 +0000 UTC m=+158.305693005 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.480565 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09726ef2-b31b-48cd-8577-7eb3ae6aaca6-utilities\") pod \"certified-operators-zbj7g\" (UID: \"09726ef2-b31b-48cd-8577-7eb3ae6aaca6\") " pod="openshift-marketplace/certified-operators-zbj7g" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.480911 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09726ef2-b31b-48cd-8577-7eb3ae6aaca6-catalog-content\") pod \"certified-operators-zbj7g\" (UID: \"09726ef2-b31b-48cd-8577-7eb3ae6aaca6\") " pod="openshift-marketplace/certified-operators-zbj7g" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.496461 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lhm6x"] Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.519718 4808 patch_prober.go:28] interesting pod/router-default-5444994796-7b9zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 16:43:10 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Oct 02 16:43:10 crc kubenswrapper[4808]: [+]process-running ok Oct 02 16:43:10 crc kubenswrapper[4808]: healthz check failed Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.520305 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7b9zg" podUID="9ebb9933-4652-4152-a529-5489fd9d5e5d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.573532 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tqpg\" (UniqueName: \"kubernetes.io/projected/09726ef2-b31b-48cd-8577-7eb3ae6aaca6-kube-api-access-6tqpg\") pod \"certified-operators-zbj7g\" (UID: \"09726ef2-b31b-48cd-8577-7eb3ae6aaca6\") " pod="openshift-marketplace/certified-operators-zbj7g" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.581598 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zbj7g" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.583626 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.583702 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/429c5d29-dd2e-4938-a8cb-08a19212a6fc-catalog-content\") pod \"community-operators-lhm6x\" (UID: \"429c5d29-dd2e-4938-a8cb-08a19212a6fc\") " pod="openshift-marketplace/community-operators-lhm6x" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.583751 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9jmk\" (UniqueName: \"kubernetes.io/projected/429c5d29-dd2e-4938-a8cb-08a19212a6fc-kube-api-access-c9jmk\") pod \"community-operators-lhm6x\" (UID: \"429c5d29-dd2e-4938-a8cb-08a19212a6fc\") " pod="openshift-marketplace/community-operators-lhm6x" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.583803 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/429c5d29-dd2e-4938-a8cb-08a19212a6fc-utilities\") pod \"community-operators-lhm6x\" (UID: \"429c5d29-dd2e-4938-a8cb-08a19212a6fc\") " pod="openshift-marketplace/community-operators-lhm6x" Oct 02 16:43:10 crc kubenswrapper[4808]: E1002 16:43:10.584416 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:11.084401265 +0000 UTC m=+158.409930265 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.605393 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kbn54"] Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.606755 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kbn54" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.637194 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kbn54"] Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.684649 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.685262 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/429c5d29-dd2e-4938-a8cb-08a19212a6fc-catalog-content\") pod \"community-operators-lhm6x\" (UID: \"429c5d29-dd2e-4938-a8cb-08a19212a6fc\") " pod="openshift-marketplace/community-operators-lhm6x" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.685365 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9jmk\" (UniqueName: \"kubernetes.io/projected/429c5d29-dd2e-4938-a8cb-08a19212a6fc-kube-api-access-c9jmk\") pod \"community-operators-lhm6x\" (UID: \"429c5d29-dd2e-4938-a8cb-08a19212a6fc\") " pod="openshift-marketplace/community-operators-lhm6x" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.685463 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/429c5d29-dd2e-4938-a8cb-08a19212a6fc-utilities\") pod \"community-operators-lhm6x\" (UID: \"429c5d29-dd2e-4938-a8cb-08a19212a6fc\") " pod="openshift-marketplace/community-operators-lhm6x" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.685989 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/429c5d29-dd2e-4938-a8cb-08a19212a6fc-utilities\") pod \"community-operators-lhm6x\" (UID: \"429c5d29-dd2e-4938-a8cb-08a19212a6fc\") " pod="openshift-marketplace/community-operators-lhm6x" Oct 02 16:43:10 crc kubenswrapper[4808]: E1002 16:43:10.686434 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:11.186403458 +0000 UTC m=+158.511932458 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.686446 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/429c5d29-dd2e-4938-a8cb-08a19212a6fc-catalog-content\") pod \"community-operators-lhm6x\" (UID: \"429c5d29-dd2e-4938-a8cb-08a19212a6fc\") " pod="openshift-marketplace/community-operators-lhm6x" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.717100 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9jmk\" (UniqueName: \"kubernetes.io/projected/429c5d29-dd2e-4938-a8cb-08a19212a6fc-kube-api-access-c9jmk\") pod \"community-operators-lhm6x\" (UID: \"429c5d29-dd2e-4938-a8cb-08a19212a6fc\") " pod="openshift-marketplace/community-operators-lhm6x" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.786608 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9439cd36-0f01-460d-a440-ac61df1bbb6f-catalog-content\") pod \"certified-operators-kbn54\" (UID: \"9439cd36-0f01-460d-a440-ac61df1bbb6f\") " pod="openshift-marketplace/certified-operators-kbn54" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.786655 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9439cd36-0f01-460d-a440-ac61df1bbb6f-utilities\") pod \"certified-operators-kbn54\" (UID: \"9439cd36-0f01-460d-a440-ac61df1bbb6f\") " pod="openshift-marketplace/certified-operators-kbn54" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.786704 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.786735 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z52f6\" (UniqueName: \"kubernetes.io/projected/9439cd36-0f01-460d-a440-ac61df1bbb6f-kube-api-access-z52f6\") pod \"certified-operators-kbn54\" (UID: \"9439cd36-0f01-460d-a440-ac61df1bbb6f\") " pod="openshift-marketplace/certified-operators-kbn54" Oct 02 16:43:10 crc kubenswrapper[4808]: E1002 16:43:10.787115 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:11.287097708 +0000 UTC m=+158.612626708 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.887948 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:10 crc kubenswrapper[4808]: E1002 16:43:10.888576 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:11.388517247 +0000 UTC m=+158.714046247 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.888650 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9439cd36-0f01-460d-a440-ac61df1bbb6f-utilities\") pod \"certified-operators-kbn54\" (UID: \"9439cd36-0f01-460d-a440-ac61df1bbb6f\") " pod="openshift-marketplace/certified-operators-kbn54" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.888706 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.888743 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z52f6\" (UniqueName: \"kubernetes.io/projected/9439cd36-0f01-460d-a440-ac61df1bbb6f-kube-api-access-z52f6\") pod \"certified-operators-kbn54\" (UID: \"9439cd36-0f01-460d-a440-ac61df1bbb6f\") " pod="openshift-marketplace/certified-operators-kbn54" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.888790 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9439cd36-0f01-460d-a440-ac61df1bbb6f-catalog-content\") pod \"certified-operators-kbn54\" (UID: \"9439cd36-0f01-460d-a440-ac61df1bbb6f\") " pod="openshift-marketplace/certified-operators-kbn54" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.889816 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9439cd36-0f01-460d-a440-ac61df1bbb6f-catalog-content\") pod \"certified-operators-kbn54\" (UID: \"9439cd36-0f01-460d-a440-ac61df1bbb6f\") " pod="openshift-marketplace/certified-operators-kbn54" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.889895 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9439cd36-0f01-460d-a440-ac61df1bbb6f-utilities\") pod \"certified-operators-kbn54\" (UID: \"9439cd36-0f01-460d-a440-ac61df1bbb6f\") " pod="openshift-marketplace/certified-operators-kbn54" Oct 02 16:43:10 crc kubenswrapper[4808]: E1002 16:43:10.890161 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:11.390151958 +0000 UTC m=+158.715680958 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.912870 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z52f6\" (UniqueName: \"kubernetes.io/projected/9439cd36-0f01-460d-a440-ac61df1bbb6f-kube-api-access-z52f6\") pod \"certified-operators-kbn54\" (UID: \"9439cd36-0f01-460d-a440-ac61df1bbb6f\") " pod="openshift-marketplace/certified-operators-kbn54" Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.961739 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zbj7g"] Oct 02 16:43:10 crc kubenswrapper[4808]: I1002 16:43:10.972338 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kbn54" Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.000737 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:11 crc kubenswrapper[4808]: E1002 16:43:11.001051 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:11.501015114 +0000 UTC m=+158.826544124 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.001208 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:11 crc kubenswrapper[4808]: E1002 16:43:11.001563 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:11.501551107 +0000 UTC m=+158.827080097 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.102051 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:11 crc kubenswrapper[4808]: E1002 16:43:11.102279 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:11.602249047 +0000 UTC m=+158.927778047 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.205921 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:11 crc kubenswrapper[4808]: E1002 16:43:11.206752 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:11.706731343 +0000 UTC m=+159.032260343 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.284673 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.292226 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nzsf5" Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.292415 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lhm6x" Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.310071 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:11 crc kubenswrapper[4808]: E1002 16:43:11.310527 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:11.810503971 +0000 UTC m=+159.136032971 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.331426 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zbj7g" event={"ID":"09726ef2-b31b-48cd-8577-7eb3ae6aaca6","Type":"ContainerStarted","Data":"dee5719c4a67f2ea73b66b7b333951b61d6ec163edb4adc147a9cf6980c16e9d"} Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.331488 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zbj7g" event={"ID":"09726ef2-b31b-48cd-8577-7eb3ae6aaca6","Type":"ContainerStarted","Data":"cb46f452e6e568465fc56dc5b372acf5ee5884a9a1a079e3d5b529866a617ad6"} Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.339007 4808 generic.go:334] "Generic (PLEG): container finished" podID="af55464e-2244-46e0-a2bc-9211829b2d7d" containerID="b7bc0a3b03df07ebd5ed98d2c6a60e3dbbebd024ce364c504ef9303785fde32f" exitCode=0 Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.339100 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw" event={"ID":"af55464e-2244-46e0-a2bc-9211829b2d7d","Type":"ContainerDied","Data":"b7bc0a3b03df07ebd5ed98d2c6a60e3dbbebd024ce364c504ef9303785fde32f"} Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.382962 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2htks" event={"ID":"681e26e4-4764-43f4-b3f9-d084f9619e5f","Type":"ContainerStarted","Data":"425d7d4e65acc852b03bcd47b5074cf9f920168836d356cc6f63e8e9ab826e0b"} Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.383384 4808 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-qc629 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.383468 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" podUID="c74723a8-81cf-4633-8655-e1460025f1b7" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.388548 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.411117 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kbn54"] Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.412535 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:11 crc kubenswrapper[4808]: E1002 16:43:11.415665 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:11.915644733 +0000 UTC m=+159.241173923 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.486346 4808 patch_prober.go:28] interesting pod/router-default-5444994796-7b9zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 16:43:11 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Oct 02 16:43:11 crc kubenswrapper[4808]: [+]process-running ok Oct 02 16:43:11 crc kubenswrapper[4808]: healthz check failed Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.486406 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7b9zg" podUID="9ebb9933-4652-4152-a529-5489fd9d5e5d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.514475 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:11 crc kubenswrapper[4808]: E1002 16:43:11.515074 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:12.01503645 +0000 UTC m=+159.340565440 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.515562 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:11 crc kubenswrapper[4808]: E1002 16:43:11.516370 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:12.016355773 +0000 UTC m=+159.341884773 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.617047 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:11 crc kubenswrapper[4808]: E1002 16:43:11.617614 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:12.117522246 +0000 UTC m=+159.443051246 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.696432 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nzsf5"] Oct 02 16:43:11 crc kubenswrapper[4808]: W1002 16:43:11.709906 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c0f4487_6602_4c9d_b414_229588b2e797.slice/crio-9428346d2a142aa8412b3b30ff6bc3094afa01b306abc0f68c150d16d2af7329 WatchSource:0}: Error finding container 9428346d2a142aa8412b3b30ff6bc3094afa01b306abc0f68c150d16d2af7329: Status 404 returned error can't find the container with id 9428346d2a142aa8412b3b30ff6bc3094afa01b306abc0f68c150d16d2af7329 Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.719251 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:11 crc kubenswrapper[4808]: E1002 16:43:11.719736 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:12.219722854 +0000 UTC m=+159.545251854 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.728893 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lhm6x"] Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.820496 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:11 crc kubenswrapper[4808]: E1002 16:43:11.820719 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:12.320686611 +0000 UTC m=+159.646215611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.820814 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:11 crc kubenswrapper[4808]: E1002 16:43:11.821192 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:12.321184773 +0000 UTC m=+159.646713773 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.883230 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.884411 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.886429 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.892642 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.894134 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.947567 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.948071 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/00ff555d-8628-48a3-b667-a2b5d1aeb249-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"00ff555d-8628-48a3-b667-a2b5d1aeb249\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 16:43:11 crc kubenswrapper[4808]: E1002 16:43:11.948632 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:12.448591534 +0000 UTC m=+159.774120524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.951123 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:11 crc kubenswrapper[4808]: I1002 16:43:11.951341 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/00ff555d-8628-48a3-b667-a2b5d1aeb249-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"00ff555d-8628-48a3-b667-a2b5d1aeb249\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 16:43:11 crc kubenswrapper[4808]: E1002 16:43:11.954695 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:12.454649616 +0000 UTC m=+159.780178616 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.053401 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:12 crc kubenswrapper[4808]: E1002 16:43:12.053789 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:12.553741936 +0000 UTC m=+159.879270946 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.053998 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/00ff555d-8628-48a3-b667-a2b5d1aeb249-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"00ff555d-8628-48a3-b667-a2b5d1aeb249\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.054141 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.054225 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/00ff555d-8628-48a3-b667-a2b5d1aeb249-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"00ff555d-8628-48a3-b667-a2b5d1aeb249\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.054161 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/00ff555d-8628-48a3-b667-a2b5d1aeb249-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"00ff555d-8628-48a3-b667-a2b5d1aeb249\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 16:43:12 crc kubenswrapper[4808]: E1002 16:43:12.054651 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:12.554627468 +0000 UTC m=+159.880156468 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.072926 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/00ff555d-8628-48a3-b667-a2b5d1aeb249-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"00ff555d-8628-48a3-b667-a2b5d1aeb249\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.155378 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:12 crc kubenswrapper[4808]: E1002 16:43:12.155662 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:12.655619996 +0000 UTC m=+159.981148996 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.156115 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:12 crc kubenswrapper[4808]: E1002 16:43:12.156682 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:12.656658712 +0000 UTC m=+159.982187702 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.200828 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zhxtz"] Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.202032 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zhxtz" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.204987 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.215619 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhxtz"] Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.268351 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.268725 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.269048 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsjsh\" (UniqueName: \"kubernetes.io/projected/fa0909cd-83b3-4127-abc6-2540f61e0b2b-kube-api-access-bsjsh\") pod \"redhat-marketplace-zhxtz\" (UID: \"fa0909cd-83b3-4127-abc6-2540f61e0b2b\") " pod="openshift-marketplace/redhat-marketplace-zhxtz" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.269164 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa0909cd-83b3-4127-abc6-2540f61e0b2b-utilities\") pod \"redhat-marketplace-zhxtz\" (UID: \"fa0909cd-83b3-4127-abc6-2540f61e0b2b\") " pod="openshift-marketplace/redhat-marketplace-zhxtz" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.269310 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa0909cd-83b3-4127-abc6-2540f61e0b2b-catalog-content\") pod \"redhat-marketplace-zhxtz\" (UID: \"fa0909cd-83b3-4127-abc6-2540f61e0b2b\") " pod="openshift-marketplace/redhat-marketplace-zhxtz" Oct 02 16:43:12 crc kubenswrapper[4808]: E1002 16:43:12.269484 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:12.769461387 +0000 UTC m=+160.094990387 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.356131 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.370719 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa0909cd-83b3-4127-abc6-2540f61e0b2b-catalog-content\") pod \"redhat-marketplace-zhxtz\" (UID: \"fa0909cd-83b3-4127-abc6-2540f61e0b2b\") " pod="openshift-marketplace/redhat-marketplace-zhxtz" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.370775 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsjsh\" (UniqueName: \"kubernetes.io/projected/fa0909cd-83b3-4127-abc6-2540f61e0b2b-kube-api-access-bsjsh\") pod \"redhat-marketplace-zhxtz\" (UID: \"fa0909cd-83b3-4127-abc6-2540f61e0b2b\") " pod="openshift-marketplace/redhat-marketplace-zhxtz" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.370816 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa0909cd-83b3-4127-abc6-2540f61e0b2b-utilities\") pod \"redhat-marketplace-zhxtz\" (UID: \"fa0909cd-83b3-4127-abc6-2540f61e0b2b\") " pod="openshift-marketplace/redhat-marketplace-zhxtz" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.370849 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:12 crc kubenswrapper[4808]: E1002 16:43:12.371182 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:12.871169893 +0000 UTC m=+160.196698893 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.372127 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa0909cd-83b3-4127-abc6-2540f61e0b2b-catalog-content\") pod \"redhat-marketplace-zhxtz\" (UID: \"fa0909cd-83b3-4127-abc6-2540f61e0b2b\") " pod="openshift-marketplace/redhat-marketplace-zhxtz" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.372386 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa0909cd-83b3-4127-abc6-2540f61e0b2b-utilities\") pod \"redhat-marketplace-zhxtz\" (UID: \"fa0909cd-83b3-4127-abc6-2540f61e0b2b\") " pod="openshift-marketplace/redhat-marketplace-zhxtz" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.413627 4808 generic.go:334] "Generic (PLEG): container finished" podID="9439cd36-0f01-460d-a440-ac61df1bbb6f" containerID="eb1d38a79c5806e30077f28e60f721d0c564bcf337b09e3b5f1b4e267d379b20" exitCode=0 Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.414127 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbn54" event={"ID":"9439cd36-0f01-460d-a440-ac61df1bbb6f","Type":"ContainerDied","Data":"eb1d38a79c5806e30077f28e60f721d0c564bcf337b09e3b5f1b4e267d379b20"} Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.414160 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbn54" event={"ID":"9439cd36-0f01-460d-a440-ac61df1bbb6f","Type":"ContainerStarted","Data":"3a259a0da166b9b5d132a46230967f9b403400de441f9134b4919047e000ef14"} Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.424160 4808 generic.go:334] "Generic (PLEG): container finished" podID="6c0f4487-6602-4c9d-b414-229588b2e797" containerID="df0b439bf0b7cac24879d695e54118304faeecc8026be631c1d497c428de9b4d" exitCode=0 Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.424226 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nzsf5" event={"ID":"6c0f4487-6602-4c9d-b414-229588b2e797","Type":"ContainerDied","Data":"df0b439bf0b7cac24879d695e54118304faeecc8026be631c1d497c428de9b4d"} Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.424279 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nzsf5" event={"ID":"6c0f4487-6602-4c9d-b414-229588b2e797","Type":"ContainerStarted","Data":"9428346d2a142aa8412b3b30ff6bc3094afa01b306abc0f68c150d16d2af7329"} Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.425664 4808 generic.go:334] "Generic (PLEG): container finished" podID="09726ef2-b31b-48cd-8577-7eb3ae6aaca6" containerID="dee5719c4a67f2ea73b66b7b333951b61d6ec163edb4adc147a9cf6980c16e9d" exitCode=0 Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.425701 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zbj7g" event={"ID":"09726ef2-b31b-48cd-8577-7eb3ae6aaca6","Type":"ContainerDied","Data":"dee5719c4a67f2ea73b66b7b333951b61d6ec163edb4adc147a9cf6980c16e9d"} Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.433160 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.445903 4808 generic.go:334] "Generic (PLEG): container finished" podID="429c5d29-dd2e-4938-a8cb-08a19212a6fc" containerID="2bba5502fe52cfc26051e1235f5122f585641a61caa7d7dc5d00ab6c3a59dba7" exitCode=0 Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.446006 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lhm6x" event={"ID":"429c5d29-dd2e-4938-a8cb-08a19212a6fc","Type":"ContainerDied","Data":"2bba5502fe52cfc26051e1235f5122f585641a61caa7d7dc5d00ab6c3a59dba7"} Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.446107 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lhm6x" event={"ID":"429c5d29-dd2e-4938-a8cb-08a19212a6fc","Type":"ContainerStarted","Data":"83fbb3eae0ed5fdd3bcea86977934130f2ea839bfc689e1d0d1af90a7699815c"} Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.454386 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsjsh\" (UniqueName: \"kubernetes.io/projected/fa0909cd-83b3-4127-abc6-2540f61e0b2b-kube-api-access-bsjsh\") pod \"redhat-marketplace-zhxtz\" (UID: \"fa0909cd-83b3-4127-abc6-2540f61e0b2b\") " pod="openshift-marketplace/redhat-marketplace-zhxtz" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.472792 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:12 crc kubenswrapper[4808]: E1002 16:43:12.474693 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:12.974670714 +0000 UTC m=+160.300199714 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.486051 4808 patch_prober.go:28] interesting pod/router-default-5444994796-7b9zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 16:43:12 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Oct 02 16:43:12 crc kubenswrapper[4808]: [+]process-running ok Oct 02 16:43:12 crc kubenswrapper[4808]: healthz check failed Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.486104 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7b9zg" podUID="9ebb9933-4652-4152-a529-5489fd9d5e5d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.517660 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zhxtz" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.575605 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:12 crc kubenswrapper[4808]: E1002 16:43:12.577344 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:13.077327393 +0000 UTC m=+160.402856393 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.613429 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-22l54"] Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.614559 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-22l54" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.625570 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-22l54"] Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.682007 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:12 crc kubenswrapper[4808]: E1002 16:43:12.683016 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:13.182970038 +0000 UTC m=+160.508499038 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.691439 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25e17c62-b490-4c5d-8411-ef484d9039fc-utilities\") pod \"redhat-marketplace-22l54\" (UID: \"25e17c62-b490-4c5d-8411-ef484d9039fc\") " pod="openshift-marketplace/redhat-marketplace-22l54" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.691600 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spv6h\" (UniqueName: \"kubernetes.io/projected/25e17c62-b490-4c5d-8411-ef484d9039fc-kube-api-access-spv6h\") pod \"redhat-marketplace-22l54\" (UID: \"25e17c62-b490-4c5d-8411-ef484d9039fc\") " pod="openshift-marketplace/redhat-marketplace-22l54" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.691758 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.691863 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25e17c62-b490-4c5d-8411-ef484d9039fc-catalog-content\") pod \"redhat-marketplace-22l54\" (UID: \"25e17c62-b490-4c5d-8411-ef484d9039fc\") " pod="openshift-marketplace/redhat-marketplace-22l54" Oct 02 16:43:12 crc kubenswrapper[4808]: E1002 16:43:12.692358 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:13.192342623 +0000 UTC m=+160.517871613 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.733874 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.793501 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdtb8\" (UniqueName: \"kubernetes.io/projected/af55464e-2244-46e0-a2bc-9211829b2d7d-kube-api-access-sdtb8\") pod \"af55464e-2244-46e0-a2bc-9211829b2d7d\" (UID: \"af55464e-2244-46e0-a2bc-9211829b2d7d\") " Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.793666 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.793694 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af55464e-2244-46e0-a2bc-9211829b2d7d-config-volume\") pod \"af55464e-2244-46e0-a2bc-9211829b2d7d\" (UID: \"af55464e-2244-46e0-a2bc-9211829b2d7d\") " Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.793729 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af55464e-2244-46e0-a2bc-9211829b2d7d-secret-volume\") pod \"af55464e-2244-46e0-a2bc-9211829b2d7d\" (UID: \"af55464e-2244-46e0-a2bc-9211829b2d7d\") " Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.793902 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25e17c62-b490-4c5d-8411-ef484d9039fc-catalog-content\") pod \"redhat-marketplace-22l54\" (UID: \"25e17c62-b490-4c5d-8411-ef484d9039fc\") " pod="openshift-marketplace/redhat-marketplace-22l54" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.793950 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25e17c62-b490-4c5d-8411-ef484d9039fc-utilities\") pod \"redhat-marketplace-22l54\" (UID: \"25e17c62-b490-4c5d-8411-ef484d9039fc\") " pod="openshift-marketplace/redhat-marketplace-22l54" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.793989 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spv6h\" (UniqueName: \"kubernetes.io/projected/25e17c62-b490-4c5d-8411-ef484d9039fc-kube-api-access-spv6h\") pod \"redhat-marketplace-22l54\" (UID: \"25e17c62-b490-4c5d-8411-ef484d9039fc\") " pod="openshift-marketplace/redhat-marketplace-22l54" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.800613 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af55464e-2244-46e0-a2bc-9211829b2d7d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "af55464e-2244-46e0-a2bc-9211829b2d7d" (UID: "af55464e-2244-46e0-a2bc-9211829b2d7d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:43:12 crc kubenswrapper[4808]: E1002 16:43:12.800744 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:13.300724427 +0000 UTC m=+160.626253427 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.802126 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25e17c62-b490-4c5d-8411-ef484d9039fc-catalog-content\") pod \"redhat-marketplace-22l54\" (UID: \"25e17c62-b490-4c5d-8411-ef484d9039fc\") " pod="openshift-marketplace/redhat-marketplace-22l54" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.802423 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25e17c62-b490-4c5d-8411-ef484d9039fc-utilities\") pod \"redhat-marketplace-22l54\" (UID: \"25e17c62-b490-4c5d-8411-ef484d9039fc\") " pod="openshift-marketplace/redhat-marketplace-22l54" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.802974 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af55464e-2244-46e0-a2bc-9211829b2d7d-kube-api-access-sdtb8" (OuterVolumeSpecName: "kube-api-access-sdtb8") pod "af55464e-2244-46e0-a2bc-9211829b2d7d" (UID: "af55464e-2244-46e0-a2bc-9211829b2d7d"). InnerVolumeSpecName "kube-api-access-sdtb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.813845 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af55464e-2244-46e0-a2bc-9211829b2d7d-config-volume" (OuterVolumeSpecName: "config-volume") pod "af55464e-2244-46e0-a2bc-9211829b2d7d" (UID: "af55464e-2244-46e0-a2bc-9211829b2d7d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.826868 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.861964 4808 patch_prober.go:28] interesting pod/downloads-7954f5f757-qsqkp container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.862028 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-qsqkp" podUID="8bc8bd65-c754-4a0a-9f11-40d444e2492a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.861964 4808 patch_prober.go:28] interesting pod/downloads-7954f5f757-qsqkp container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.862406 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qsqkp" podUID="8bc8bd65-c754-4a0a-9f11-40d444e2492a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.870082 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spv6h\" (UniqueName: \"kubernetes.io/projected/25e17c62-b490-4c5d-8411-ef484d9039fc-kube-api-access-spv6h\") pod \"redhat-marketplace-22l54\" (UID: \"25e17c62-b490-4c5d-8411-ef484d9039fc\") " pod="openshift-marketplace/redhat-marketplace-22l54" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.900516 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.900580 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdtb8\" (UniqueName: \"kubernetes.io/projected/af55464e-2244-46e0-a2bc-9211829b2d7d-kube-api-access-sdtb8\") on node \"crc\" DevicePath \"\"" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.900591 4808 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af55464e-2244-46e0-a2bc-9211829b2d7d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.900602 4808 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af55464e-2244-46e0-a2bc-9211829b2d7d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 16:43:12 crc kubenswrapper[4808]: E1002 16:43:12.900925 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:13.400910784 +0000 UTC m=+160.726439784 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.932142 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhxtz"] Oct 02 16:43:12 crc kubenswrapper[4808]: W1002 16:43:12.950969 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa0909cd_83b3_4127_abc6_2540f61e0b2b.slice/crio-93b68e787fb922d50207f8ee2c5cac6ee72af1f3a9790354ed4cd09a42d0a668 WatchSource:0}: Error finding container 93b68e787fb922d50207f8ee2c5cac6ee72af1f3a9790354ed4cd09a42d0a668: Status 404 returned error can't find the container with id 93b68e787fb922d50207f8ee2c5cac6ee72af1f3a9790354ed4cd09a42d0a668 Oct 02 16:43:12 crc kubenswrapper[4808]: I1002 16:43:12.956679 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-22l54" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.001823 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:13 crc kubenswrapper[4808]: E1002 16:43:13.002063 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:13.502027145 +0000 UTC m=+160.827556135 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.104443 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:13 crc kubenswrapper[4808]: E1002 16:43:13.104914 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:13.60489285 +0000 UTC m=+160.930421850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.208214 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:13 crc kubenswrapper[4808]: E1002 16:43:13.208745 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:13.708712769 +0000 UTC m=+161.034241769 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.211202 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dz6nh"] Oct 02 16:43:13 crc kubenswrapper[4808]: E1002 16:43:13.218540 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af55464e-2244-46e0-a2bc-9211829b2d7d" containerName="collect-profiles" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.218563 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="af55464e-2244-46e0-a2bc-9211829b2d7d" containerName="collect-profiles" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.218701 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="af55464e-2244-46e0-a2bc-9211829b2d7d" containerName="collect-profiles" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.227591 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dz6nh" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.235704 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.236114 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dz6nh"] Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.310094 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.310147 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21467289-5a33-43ed-ac9b-425774019de0-catalog-content\") pod \"redhat-operators-dz6nh\" (UID: \"21467289-5a33-43ed-ac9b-425774019de0\") " pod="openshift-marketplace/redhat-operators-dz6nh" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.310173 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21467289-5a33-43ed-ac9b-425774019de0-utilities\") pod \"redhat-operators-dz6nh\" (UID: \"21467289-5a33-43ed-ac9b-425774019de0\") " pod="openshift-marketplace/redhat-operators-dz6nh" Oct 02 16:43:13 crc kubenswrapper[4808]: E1002 16:43:13.311552 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:13.811535463 +0000 UTC m=+161.137064463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.312269 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr54f\" (UniqueName: \"kubernetes.io/projected/21467289-5a33-43ed-ac9b-425774019de0-kube-api-access-rr54f\") pod \"redhat-operators-dz6nh\" (UID: \"21467289-5a33-43ed-ac9b-425774019de0\") " pod="openshift-marketplace/redhat-operators-dz6nh" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.413884 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.414150 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr54f\" (UniqueName: \"kubernetes.io/projected/21467289-5a33-43ed-ac9b-425774019de0-kube-api-access-rr54f\") pod \"redhat-operators-dz6nh\" (UID: \"21467289-5a33-43ed-ac9b-425774019de0\") " pod="openshift-marketplace/redhat-operators-dz6nh" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.414227 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21467289-5a33-43ed-ac9b-425774019de0-catalog-content\") pod \"redhat-operators-dz6nh\" (UID: \"21467289-5a33-43ed-ac9b-425774019de0\") " pod="openshift-marketplace/redhat-operators-dz6nh" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.414274 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21467289-5a33-43ed-ac9b-425774019de0-utilities\") pod \"redhat-operators-dz6nh\" (UID: \"21467289-5a33-43ed-ac9b-425774019de0\") " pod="openshift-marketplace/redhat-operators-dz6nh" Oct 02 16:43:13 crc kubenswrapper[4808]: E1002 16:43:13.414590 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:13.914553852 +0000 UTC m=+161.240082852 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.414937 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21467289-5a33-43ed-ac9b-425774019de0-utilities\") pod \"redhat-operators-dz6nh\" (UID: \"21467289-5a33-43ed-ac9b-425774019de0\") " pod="openshift-marketplace/redhat-operators-dz6nh" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.416476 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21467289-5a33-43ed-ac9b-425774019de0-catalog-content\") pod \"redhat-operators-dz6nh\" (UID: \"21467289-5a33-43ed-ac9b-425774019de0\") " pod="openshift-marketplace/redhat-operators-dz6nh" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.442249 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr54f\" (UniqueName: \"kubernetes.io/projected/21467289-5a33-43ed-ac9b-425774019de0-kube-api-access-rr54f\") pod \"redhat-operators-dz6nh\" (UID: \"21467289-5a33-43ed-ac9b-425774019de0\") " pod="openshift-marketplace/redhat-operators-dz6nh" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.474523 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"00ff555d-8628-48a3-b667-a2b5d1aeb249","Type":"ContainerStarted","Data":"fd570a49b3a348218124f63c34e0a7548cb33ef91284fb85f3b5f9962d483445"} Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.475004 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"00ff555d-8628-48a3-b667-a2b5d1aeb249","Type":"ContainerStarted","Data":"7ab68f660811672024a1624b38a5bc77329f1f0b38e6fe0fbde18089db41f99e"} Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.477164 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.477398 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323710-kllgw" event={"ID":"af55464e-2244-46e0-a2bc-9211829b2d7d","Type":"ContainerDied","Data":"9ed3e509e7b8032cf79b496b373898d86249af6a29804e8770ddf820d3277589"} Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.477478 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ed3e509e7b8032cf79b496b373898d86249af6a29804e8770ddf820d3277589" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.485467 4808 generic.go:334] "Generic (PLEG): container finished" podID="fa0909cd-83b3-4127-abc6-2540f61e0b2b" containerID="5ad2f3ab411e52716530e2532e0b72602313552c4427d5baf112290ca04ea400" exitCode=0 Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.485521 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhxtz" event={"ID":"fa0909cd-83b3-4127-abc6-2540f61e0b2b","Type":"ContainerDied","Data":"5ad2f3ab411e52716530e2532e0b72602313552c4427d5baf112290ca04ea400"} Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.485555 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhxtz" event={"ID":"fa0909cd-83b3-4127-abc6-2540f61e0b2b","Type":"ContainerStarted","Data":"93b68e787fb922d50207f8ee2c5cac6ee72af1f3a9790354ed4cd09a42d0a668"} Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.493665 4808 patch_prober.go:28] interesting pod/router-default-5444994796-7b9zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 16:43:13 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Oct 02 16:43:13 crc kubenswrapper[4808]: [+]process-running ok Oct 02 16:43:13 crc kubenswrapper[4808]: healthz check failed Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.493740 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7b9zg" podUID="9ebb9933-4652-4152-a529-5489fd9d5e5d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.519159 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:13 crc kubenswrapper[4808]: E1002 16:43:13.519576 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:14.01956179 +0000 UTC m=+161.345090790 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.556394 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-22l54"] Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.599419 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wwrlz"] Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.602844 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wwrlz" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.617749 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wwrlz"] Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.621000 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:13 crc kubenswrapper[4808]: E1002 16:43:13.621758 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:14.121726978 +0000 UTC m=+161.447255978 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.645916 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dz6nh" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.724784 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.725138 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsszr\" (UniqueName: \"kubernetes.io/projected/d621b4b0-80ad-489e-8f63-76918632c5ef-kube-api-access-vsszr\") pod \"redhat-operators-wwrlz\" (UID: \"d621b4b0-80ad-489e-8f63-76918632c5ef\") " pod="openshift-marketplace/redhat-operators-wwrlz" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.725271 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d621b4b0-80ad-489e-8f63-76918632c5ef-utilities\") pod \"redhat-operators-wwrlz\" (UID: \"d621b4b0-80ad-489e-8f63-76918632c5ef\") " pod="openshift-marketplace/redhat-operators-wwrlz" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.725381 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d621b4b0-80ad-489e-8f63-76918632c5ef-catalog-content\") pod \"redhat-operators-wwrlz\" (UID: \"d621b4b0-80ad-489e-8f63-76918632c5ef\") " pod="openshift-marketplace/redhat-operators-wwrlz" Oct 02 16:43:13 crc kubenswrapper[4808]: E1002 16:43:13.725414 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:14.225388062 +0000 UTC m=+161.550917182 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.827901 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.828606 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsszr\" (UniqueName: \"kubernetes.io/projected/d621b4b0-80ad-489e-8f63-76918632c5ef-kube-api-access-vsszr\") pod \"redhat-operators-wwrlz\" (UID: \"d621b4b0-80ad-489e-8f63-76918632c5ef\") " pod="openshift-marketplace/redhat-operators-wwrlz" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.828646 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d621b4b0-80ad-489e-8f63-76918632c5ef-utilities\") pod \"redhat-operators-wwrlz\" (UID: \"d621b4b0-80ad-489e-8f63-76918632c5ef\") " pod="openshift-marketplace/redhat-operators-wwrlz" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.828680 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d621b4b0-80ad-489e-8f63-76918632c5ef-catalog-content\") pod \"redhat-operators-wwrlz\" (UID: \"d621b4b0-80ad-489e-8f63-76918632c5ef\") " pod="openshift-marketplace/redhat-operators-wwrlz" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.829566 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d621b4b0-80ad-489e-8f63-76918632c5ef-catalog-content\") pod \"redhat-operators-wwrlz\" (UID: \"d621b4b0-80ad-489e-8f63-76918632c5ef\") " pod="openshift-marketplace/redhat-operators-wwrlz" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.829803 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d621b4b0-80ad-489e-8f63-76918632c5ef-utilities\") pod \"redhat-operators-wwrlz\" (UID: \"d621b4b0-80ad-489e-8f63-76918632c5ef\") " pod="openshift-marketplace/redhat-operators-wwrlz" Oct 02 16:43:13 crc kubenswrapper[4808]: E1002 16:43:13.829882 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:14.329861488 +0000 UTC m=+161.655390478 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.875089 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsszr\" (UniqueName: \"kubernetes.io/projected/d621b4b0-80ad-489e-8f63-76918632c5ef-kube-api-access-vsszr\") pod \"redhat-operators-wwrlz\" (UID: \"d621b4b0-80ad-489e-8f63-76918632c5ef\") " pod="openshift-marketplace/redhat-operators-wwrlz" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.928577 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wwrlz" Oct 02 16:43:13 crc kubenswrapper[4808]: I1002 16:43:13.929655 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:13 crc kubenswrapper[4808]: E1002 16:43:13.930561 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:14.430545628 +0000 UTC m=+161.756074628 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.001006 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dz6nh"] Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.031101 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:14 crc kubenswrapper[4808]: E1002 16:43:14.031465 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:14.531441683 +0000 UTC m=+161.856970683 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.133255 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:14 crc kubenswrapper[4808]: E1002 16:43:14.133701 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:14.633684452 +0000 UTC m=+161.959213452 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.234293 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:14 crc kubenswrapper[4808]: E1002 16:43:14.236709 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:14.73666161 +0000 UTC m=+162.062190610 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.261972 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.262012 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.274375 4808 patch_prober.go:28] interesting pod/console-f9d7485db-dlddx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.274462 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-dlddx" podUID="c58d3855-bb32-4bb3-94ae-53c49209655e" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.288728 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.294617 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-2bgkb" Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.328905 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.339521 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:14 crc kubenswrapper[4808]: E1002 16:43:14.345373 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:14.845347111 +0000 UTC m=+162.170876111 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.357450 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.365336 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wvmf" Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.441433 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:14 crc kubenswrapper[4808]: E1002 16:43:14.442554 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:14.942518673 +0000 UTC m=+162.268047833 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.478204 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.499144 4808 patch_prober.go:28] interesting pod/router-default-5444994796-7b9zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 16:43:14 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Oct 02 16:43:14 crc kubenswrapper[4808]: [+]process-running ok Oct 02 16:43:14 crc kubenswrapper[4808]: healthz check failed Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.499207 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7b9zg" podUID="9ebb9933-4652-4152-a529-5489fd9d5e5d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.544385 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:14 crc kubenswrapper[4808]: E1002 16:43:14.544772 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:15.044757682 +0000 UTC m=+162.370286682 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.573545 4808 generic.go:334] "Generic (PLEG): container finished" podID="00ff555d-8628-48a3-b667-a2b5d1aeb249" containerID="fd570a49b3a348218124f63c34e0a7548cb33ef91284fb85f3b5f9962d483445" exitCode=0 Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.573635 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"00ff555d-8628-48a3-b667-a2b5d1aeb249","Type":"ContainerDied","Data":"fd570a49b3a348218124f63c34e0a7548cb33ef91284fb85f3b5f9962d483445"} Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.575446 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dz6nh" event={"ID":"21467289-5a33-43ed-ac9b-425774019de0","Type":"ContainerStarted","Data":"1f347099fc8b55be1ed5172e26616a8e60afbdd2547dd27e07011f3999bab1d9"} Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.576462 4808 generic.go:334] "Generic (PLEG): container finished" podID="25e17c62-b490-4c5d-8411-ef484d9039fc" containerID="f47657a5c62497b5768bd4f0d8ea00c5770c779e63e445b7d4e40bd18bd8f291" exitCode=0 Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.577391 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-22l54" event={"ID":"25e17c62-b490-4c5d-8411-ef484d9039fc","Type":"ContainerDied","Data":"f47657a5c62497b5768bd4f0d8ea00c5770c779e63e445b7d4e40bd18bd8f291"} Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.577415 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-22l54" event={"ID":"25e17c62-b490-4c5d-8411-ef484d9039fc","Type":"ContainerStarted","Data":"39a7dff4a485fa17cfd97a19a5e6b645de548f1618b2b7f7926ebb7bcb90710a"} Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.642590 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wwrlz"] Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.647947 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:14 crc kubenswrapper[4808]: E1002 16:43:14.648071 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:15.148050388 +0000 UTC m=+162.473579388 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.650077 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:14 crc kubenswrapper[4808]: E1002 16:43:14.650398 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:15.150382686 +0000 UTC m=+162.475911686 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.754643 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:14 crc kubenswrapper[4808]: E1002 16:43:14.755534 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:15.255509068 +0000 UTC m=+162.581038068 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.858717 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:14 crc kubenswrapper[4808]: E1002 16:43:14.859223 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:15.359199934 +0000 UTC m=+162.684728924 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.896225 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.897326 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.903336 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.913109 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.931474 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.959976 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:14 crc kubenswrapper[4808]: E1002 16:43:14.960112 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:15.460078949 +0000 UTC m=+162.785607949 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.960255 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f6bc16bd-85c9-4ec5-b636-38b69f616aa8-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f6bc16bd-85c9-4ec5-b636-38b69f616aa8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.960290 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6bc16bd-85c9-4ec5-b636-38b69f616aa8-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f6bc16bd-85c9-4ec5-b636-38b69f616aa8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 16:43:14 crc kubenswrapper[4808]: I1002 16:43:14.960326 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:14 crc kubenswrapper[4808]: E1002 16:43:14.960800 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:15.460783796 +0000 UTC m=+162.786312796 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.063863 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:15 crc kubenswrapper[4808]: E1002 16:43:15.064130 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:15.564093892 +0000 UTC m=+162.889622882 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.064192 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f6bc16bd-85c9-4ec5-b636-38b69f616aa8-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f6bc16bd-85c9-4ec5-b636-38b69f616aa8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.064247 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6bc16bd-85c9-4ec5-b636-38b69f616aa8-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f6bc16bd-85c9-4ec5-b636-38b69f616aa8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.064284 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.064338 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f6bc16bd-85c9-4ec5-b636-38b69f616aa8-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f6bc16bd-85c9-4ec5-b636-38b69f616aa8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 16:43:15 crc kubenswrapper[4808]: E1002 16:43:15.064645 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:15.564630856 +0000 UTC m=+162.890160056 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.106125 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6bc16bd-85c9-4ec5-b636-38b69f616aa8-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f6bc16bd-85c9-4ec5-b636-38b69f616aa8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.167194 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:15 crc kubenswrapper[4808]: E1002 16:43:15.167496 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:15.66747391 +0000 UTC m=+162.993002910 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.167613 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:15 crc kubenswrapper[4808]: E1002 16:43:15.167973 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:15.667961142 +0000 UTC m=+162.993490142 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.222491 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.227124 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.273879 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:15 crc kubenswrapper[4808]: E1002 16:43:15.274300 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:15.774273634 +0000 UTC m=+163.099802814 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.294389 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dp2fq" Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.380441 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:15 crc kubenswrapper[4808]: E1002 16:43:15.381779 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:15.881756305 +0000 UTC m=+163.207285305 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.479354 4808 patch_prober.go:28] interesting pod/router-default-5444994796-7b9zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 16:43:15 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Oct 02 16:43:15 crc kubenswrapper[4808]: [+]process-running ok Oct 02 16:43:15 crc kubenswrapper[4808]: healthz check failed Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.479422 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7b9zg" podUID="9ebb9933-4652-4152-a529-5489fd9d5e5d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.483513 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:15 crc kubenswrapper[4808]: E1002 16:43:15.485121 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:15.985094841 +0000 UTC m=+163.310623841 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.585136 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:15 crc kubenswrapper[4808]: E1002 16:43:15.585663 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:16.085643007 +0000 UTC m=+163.411172007 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.605823 4808 generic.go:334] "Generic (PLEG): container finished" podID="21467289-5a33-43ed-ac9b-425774019de0" containerID="5e26b51042540c6addba14edd250706f5f348432b9ef6566b02aa7ac81c2fd97" exitCode=0 Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.605975 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dz6nh" event={"ID":"21467289-5a33-43ed-ac9b-425774019de0","Type":"ContainerDied","Data":"5e26b51042540c6addba14edd250706f5f348432b9ef6566b02aa7ac81c2fd97"} Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.629689 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2htks" event={"ID":"681e26e4-4764-43f4-b3f9-d084f9619e5f","Type":"ContainerStarted","Data":"e18d216f20f78ae11f0b66d6aa4ef12a45fe0a9ace64e9c69f535636566d8637"} Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.632545 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwrlz" event={"ID":"d621b4b0-80ad-489e-8f63-76918632c5ef","Type":"ContainerStarted","Data":"bc11c5d52a0b7b4708a38ac1819b3803904741fa77edb1af52b1e5085ee102a0"} Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.691295 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:15 crc kubenswrapper[4808]: E1002 16:43:15.691816 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:16.191739463 +0000 UTC m=+163.517268463 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.712088 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.793686 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:15 crc kubenswrapper[4808]: E1002 16:43:15.794530 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:16.294501216 +0000 UTC m=+163.620030216 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.895436 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:15 crc kubenswrapper[4808]: E1002 16:43:15.896030 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:16.396002526 +0000 UTC m=+163.721531526 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:15 crc kubenswrapper[4808]: I1002 16:43:15.997564 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:15 crc kubenswrapper[4808]: E1002 16:43:15.998190 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:16.498112562 +0000 UTC m=+163.823641562 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.036176 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.100120 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/00ff555d-8628-48a3-b667-a2b5d1aeb249-kube-api-access\") pod \"00ff555d-8628-48a3-b667-a2b5d1aeb249\" (UID: \"00ff555d-8628-48a3-b667-a2b5d1aeb249\") " Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.100187 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/00ff555d-8628-48a3-b667-a2b5d1aeb249-kubelet-dir\") pod \"00ff555d-8628-48a3-b667-a2b5d1aeb249\" (UID: \"00ff555d-8628-48a3-b667-a2b5d1aeb249\") " Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.100616 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:16 crc kubenswrapper[4808]: E1002 16:43:16.101101 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:16.60107785 +0000 UTC m=+163.926606850 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.107269 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00ff555d-8628-48a3-b667-a2b5d1aeb249-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "00ff555d-8628-48a3-b667-a2b5d1aeb249" (UID: "00ff555d-8628-48a3-b667-a2b5d1aeb249"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.107915 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00ff555d-8628-48a3-b667-a2b5d1aeb249-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "00ff555d-8628-48a3-b667-a2b5d1aeb249" (UID: "00ff555d-8628-48a3-b667-a2b5d1aeb249"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.202742 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:16 crc kubenswrapper[4808]: E1002 16:43:16.203133 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:16.703116854 +0000 UTC m=+164.028645854 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.205846 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/00ff555d-8628-48a3-b667-a2b5d1aeb249-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.205891 4808 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/00ff555d-8628-48a3-b667-a2b5d1aeb249-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.307284 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:16 crc kubenswrapper[4808]: E1002 16:43:16.307792 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:16.807767023 +0000 UTC m=+164.133296023 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.307894 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:16 crc kubenswrapper[4808]: E1002 16:43:16.308442 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:16.80841676 +0000 UTC m=+164.133945930 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.408754 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.409021 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs\") pod \"network-metrics-daemon-2x552\" (UID: \"e5e40838-d098-402e-b99a-819ce5a5977c\") " pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:43:16 crc kubenswrapper[4808]: E1002 16:43:16.409193 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:16.909150711 +0000 UTC m=+164.234679701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.429058 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5e40838-d098-402e-b99a-819ce5a5977c-metrics-certs\") pod \"network-metrics-daemon-2x552\" (UID: \"e5e40838-d098-402e-b99a-819ce5a5977c\") " pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.481803 4808 patch_prober.go:28] interesting pod/router-default-5444994796-7b9zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 16:43:16 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Oct 02 16:43:16 crc kubenswrapper[4808]: [+]process-running ok Oct 02 16:43:16 crc kubenswrapper[4808]: healthz check failed Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.481916 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7b9zg" podUID="9ebb9933-4652-4152-a529-5489fd9d5e5d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.510536 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:16 crc kubenswrapper[4808]: E1002 16:43:16.510946 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:17.010930869 +0000 UTC m=+164.336459869 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.612216 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:16 crc kubenswrapper[4808]: E1002 16:43:16.612423 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:17.112386738 +0000 UTC m=+164.437915738 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.612638 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:16 crc kubenswrapper[4808]: E1002 16:43:16.613055 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:17.113039525 +0000 UTC m=+164.438568525 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.637305 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2x552" Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.652701 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f6bc16bd-85c9-4ec5-b636-38b69f616aa8","Type":"ContainerStarted","Data":"4b9eed3589fecc7f662df7a66d0e352b44a7d956645dcd7cbabdd8c9630814b7"} Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.655606 4808 generic.go:334] "Generic (PLEG): container finished" podID="d621b4b0-80ad-489e-8f63-76918632c5ef" containerID="3c6e9eb437a4c7e14d2595688255f944212621b20ad28d92717125e9dbde0170" exitCode=0 Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.655710 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwrlz" event={"ID":"d621b4b0-80ad-489e-8f63-76918632c5ef","Type":"ContainerDied","Data":"3c6e9eb437a4c7e14d2595688255f944212621b20ad28d92717125e9dbde0170"} Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.659943 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"00ff555d-8628-48a3-b667-a2b5d1aeb249","Type":"ContainerDied","Data":"7ab68f660811672024a1624b38a5bc77329f1f0b38e6fe0fbde18089db41f99e"} Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.660004 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.660022 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ab68f660811672024a1624b38a5bc77329f1f0b38e6fe0fbde18089db41f99e" Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.715858 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:16 crc kubenswrapper[4808]: E1002 16:43:16.716002 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:17.215973641 +0000 UTC m=+164.541502641 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.716881 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:16 crc kubenswrapper[4808]: E1002 16:43:16.717258 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:17.217227003 +0000 UTC m=+164.542756003 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.818416 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:16 crc kubenswrapper[4808]: E1002 16:43:16.819344 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:17.319300948 +0000 UTC m=+164.644829958 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.921590 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-2x552"] Oct 02 16:43:16 crc kubenswrapper[4808]: I1002 16:43:16.923576 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:16 crc kubenswrapper[4808]: E1002 16:43:16.923986 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:17.423968978 +0000 UTC m=+164.749497978 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:16 crc kubenswrapper[4808]: W1002 16:43:16.952781 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5e40838_d098_402e_b99a_819ce5a5977c.slice/crio-ccef256a77423fbe747cbc89126dabcbe4cfc04585d9a07f63baa3ac47c3ec37 WatchSource:0}: Error finding container ccef256a77423fbe747cbc89126dabcbe4cfc04585d9a07f63baa3ac47c3ec37: Status 404 returned error can't find the container with id ccef256a77423fbe747cbc89126dabcbe4cfc04585d9a07f63baa3ac47c3ec37 Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.042955 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:17 crc kubenswrapper[4808]: E1002 16:43:17.043224 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:17.543193514 +0000 UTC m=+164.868722514 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.043324 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:17 crc kubenswrapper[4808]: E1002 16:43:17.043710 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:17.543695566 +0000 UTC m=+164.869224566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.145179 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:17 crc kubenswrapper[4808]: E1002 16:43:17.145461 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:17.645424633 +0000 UTC m=+164.970953633 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.145605 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:17 crc kubenswrapper[4808]: E1002 16:43:17.146105 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:17.646087219 +0000 UTC m=+164.971616219 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.249075 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:17 crc kubenswrapper[4808]: E1002 16:43:17.249416 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:17.749384775 +0000 UTC m=+165.074913775 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.249607 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:17 crc kubenswrapper[4808]: E1002 16:43:17.250155 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:17.750134364 +0000 UTC m=+165.075663364 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.351076 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:17 crc kubenswrapper[4808]: E1002 16:43:17.351258 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:17.851213424 +0000 UTC m=+165.176742424 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.351569 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:17 crc kubenswrapper[4808]: E1002 16:43:17.351984 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:17.851976713 +0000 UTC m=+165.177505713 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.453738 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:17 crc kubenswrapper[4808]: E1002 16:43:17.454107 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:17.954082309 +0000 UTC m=+165.279611299 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.454573 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:17 crc kubenswrapper[4808]: E1002 16:43:17.454976 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:17.954964661 +0000 UTC m=+165.280493661 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.481546 4808 patch_prober.go:28] interesting pod/router-default-5444994796-7b9zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 16:43:17 crc kubenswrapper[4808]: [+]has-synced ok Oct 02 16:43:17 crc kubenswrapper[4808]: [+]process-running ok Oct 02 16:43:17 crc kubenswrapper[4808]: healthz check failed Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.481638 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7b9zg" podUID="9ebb9933-4652-4152-a529-5489fd9d5e5d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.555765 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:17 crc kubenswrapper[4808]: E1002 16:43:17.555997 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:18.055972759 +0000 UTC m=+165.381501759 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.556032 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:17 crc kubenswrapper[4808]: E1002 16:43:17.556433 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:18.05642277 +0000 UTC m=+165.381951770 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:17 crc kubenswrapper[4808]: E1002 16:43:17.657208 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:18.157177322 +0000 UTC m=+165.482706312 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.658334 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.658879 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:17 crc kubenswrapper[4808]: E1002 16:43:17.659483 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:18.159443729 +0000 UTC m=+165.484972729 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.675666 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f6bc16bd-85c9-4ec5-b636-38b69f616aa8","Type":"ContainerStarted","Data":"f4415d20214bab830c9e65c7e0e7614690bac249e6858c21fbaf62e37c9e0a37"} Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.687473 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2htks" event={"ID":"681e26e4-4764-43f4-b3f9-d084f9619e5f","Type":"ContainerStarted","Data":"a405fa17b1ee015daf1c8cd86845db0a897ee70029d8298a28afcc60a146e679"} Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.703972 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2x552" event={"ID":"e5e40838-d098-402e-b99a-819ce5a5977c","Type":"ContainerStarted","Data":"ccef256a77423fbe747cbc89126dabcbe4cfc04585d9a07f63baa3ac47c3ec37"} Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.713403 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.713366164 podStartE2EDuration="3.713366164s" podCreationTimestamp="2025-10-02 16:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:17.710624145 +0000 UTC m=+165.036153165" watchObservedRunningTime="2025-10-02 16:43:17.713366164 +0000 UTC m=+165.038895164" Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.761876 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:17 crc kubenswrapper[4808]: E1002 16:43:17.762025 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:18.261986356 +0000 UTC m=+165.587515356 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.762489 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:17 crc kubenswrapper[4808]: E1002 16:43:17.762862 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:18.262847488 +0000 UTC m=+165.588376488 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.864809 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:17 crc kubenswrapper[4808]: E1002 16:43:17.865415 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:18.365390694 +0000 UTC m=+165.690919694 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:17 crc kubenswrapper[4808]: I1002 16:43:17.967057 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:17 crc kubenswrapper[4808]: E1002 16:43:17.967621 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:18.467592913 +0000 UTC m=+165.793121913 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.070973 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:18 crc kubenswrapper[4808]: E1002 16:43:18.071133 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:18.571107844 +0000 UTC m=+165.896636844 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.071727 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:18 crc kubenswrapper[4808]: E1002 16:43:18.072145 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:18.57213676 +0000 UTC m=+165.897665760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.173433 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:18 crc kubenswrapper[4808]: E1002 16:43:18.173640 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:18.673606359 +0000 UTC m=+165.999135359 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.174380 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:18 crc kubenswrapper[4808]: E1002 16:43:18.174807 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:18.674794829 +0000 UTC m=+166.000323829 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.275965 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:18 crc kubenswrapper[4808]: E1002 16:43:18.276278 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:18.776220588 +0000 UTC m=+166.101749588 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.276595 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:18 crc kubenswrapper[4808]: E1002 16:43:18.277016 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:18.777005488 +0000 UTC m=+166.102534488 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.377900 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:18 crc kubenswrapper[4808]: E1002 16:43:18.378075 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:18.878051657 +0000 UTC m=+166.203580657 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.378263 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:18 crc kubenswrapper[4808]: E1002 16:43:18.378606 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:18.878596161 +0000 UTC m=+166.204125161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.480445 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.482510 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:18 crc kubenswrapper[4808]: E1002 16:43:18.482983 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 16:43:18.982958593 +0000 UTC m=+166.308487623 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.483905 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-7b9zg" Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.504759 4808 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.584929 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:18 crc kubenswrapper[4808]: E1002 16:43:18.586443 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 16:43:19.086422833 +0000 UTC m=+166.411951883 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9vv2b" (UID: "71aafdd1-5917-4764-9503-4c266977f251") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.589446 4808 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-02T16:43:18.504803722Z","Handler":null,"Name":""} Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.616673 4808 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.616723 4808 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.686769 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.695909 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.715949 4808 generic.go:334] "Generic (PLEG): container finished" podID="f6bc16bd-85c9-4ec5-b636-38b69f616aa8" containerID="f4415d20214bab830c9e65c7e0e7614690bac249e6858c21fbaf62e37c9e0a37" exitCode=0 Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.717009 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f6bc16bd-85c9-4ec5-b636-38b69f616aa8","Type":"ContainerDied","Data":"f4415d20214bab830c9e65c7e0e7614690bac249e6858c21fbaf62e37c9e0a37"} Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.721712 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2htks" event={"ID":"681e26e4-4764-43f4-b3f9-d084f9619e5f","Type":"ContainerStarted","Data":"0c1a3281cbed950704086c93d74a915a2eadb862653992aa6f608a7456b94b27"} Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.727967 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2x552" event={"ID":"e5e40838-d098-402e-b99a-819ce5a5977c","Type":"ContainerStarted","Data":"981cb817a393f2df9735727e44b446b0f506215d21d3aca01d4494eac9d7082f"} Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.740372 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.740448 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.755738 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-2htks" podStartSLOduration=17.755716327000002 podStartE2EDuration="17.755716327s" podCreationTimestamp="2025-10-02 16:43:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:18.750128557 +0000 UTC m=+166.075657557" watchObservedRunningTime="2025-10-02 16:43:18.755716327 +0000 UTC m=+166.081245317" Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.788876 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.812668 4808 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.812723 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:18 crc kubenswrapper[4808]: I1002 16:43:18.975423 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9vv2b\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:19 crc kubenswrapper[4808]: I1002 16:43:19.235067 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:19 crc kubenswrapper[4808]: I1002 16:43:19.433132 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 02 16:43:19 crc kubenswrapper[4808]: I1002 16:43:19.748445 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2x552" event={"ID":"e5e40838-d098-402e-b99a-819ce5a5977c","Type":"ContainerStarted","Data":"00570f51f7531744bc671bef9953a5ee1ca4a9814531123bd61ebd97634124af"} Oct 02 16:43:19 crc kubenswrapper[4808]: I1002 16:43:19.776253 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-2x552" podStartSLOduration=145.77621115 podStartE2EDuration="2m25.77621115s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:19.767564833 +0000 UTC m=+167.093093853" watchObservedRunningTime="2025-10-02 16:43:19.77621115 +0000 UTC m=+167.101740150" Oct 02 16:43:20 crc kubenswrapper[4808]: I1002 16:43:20.064927 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-vh4gp" Oct 02 16:43:22 crc kubenswrapper[4808]: I1002 16:43:22.855750 4808 patch_prober.go:28] interesting pod/downloads-7954f5f757-qsqkp container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Oct 02 16:43:22 crc kubenswrapper[4808]: I1002 16:43:22.856131 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-qsqkp" podUID="8bc8bd65-c754-4a0a-9f11-40d444e2492a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Oct 02 16:43:22 crc kubenswrapper[4808]: I1002 16:43:22.855834 4808 patch_prober.go:28] interesting pod/downloads-7954f5f757-qsqkp container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Oct 02 16:43:22 crc kubenswrapper[4808]: I1002 16:43:22.856211 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qsqkp" podUID="8bc8bd65-c754-4a0a-9f11-40d444e2492a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Oct 02 16:43:24 crc kubenswrapper[4808]: I1002 16:43:24.332709 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:24 crc kubenswrapper[4808]: I1002 16:43:24.342447 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:43:32 crc kubenswrapper[4808]: I1002 16:43:32.882959 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-qsqkp" Oct 02 16:43:35 crc kubenswrapper[4808]: I1002 16:43:35.204558 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 16:43:35 crc kubenswrapper[4808]: I1002 16:43:35.301458 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f6bc16bd-85c9-4ec5-b636-38b69f616aa8-kubelet-dir\") pod \"f6bc16bd-85c9-4ec5-b636-38b69f616aa8\" (UID: \"f6bc16bd-85c9-4ec5-b636-38b69f616aa8\") " Oct 02 16:43:35 crc kubenswrapper[4808]: I1002 16:43:35.301624 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f6bc16bd-85c9-4ec5-b636-38b69f616aa8-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f6bc16bd-85c9-4ec5-b636-38b69f616aa8" (UID: "f6bc16bd-85c9-4ec5-b636-38b69f616aa8"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:43:35 crc kubenswrapper[4808]: I1002 16:43:35.301669 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6bc16bd-85c9-4ec5-b636-38b69f616aa8-kube-api-access\") pod \"f6bc16bd-85c9-4ec5-b636-38b69f616aa8\" (UID: \"f6bc16bd-85c9-4ec5-b636-38b69f616aa8\") " Oct 02 16:43:35 crc kubenswrapper[4808]: I1002 16:43:35.302095 4808 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f6bc16bd-85c9-4ec5-b636-38b69f616aa8-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 02 16:43:35 crc kubenswrapper[4808]: I1002 16:43:35.308628 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6bc16bd-85c9-4ec5-b636-38b69f616aa8-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f6bc16bd-85c9-4ec5-b636-38b69f616aa8" (UID: "f6bc16bd-85c9-4ec5-b636-38b69f616aa8"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:43:35 crc kubenswrapper[4808]: I1002 16:43:35.403722 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6bc16bd-85c9-4ec5-b636-38b69f616aa8-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 16:43:35 crc kubenswrapper[4808]: I1002 16:43:35.861103 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f6bc16bd-85c9-4ec5-b636-38b69f616aa8","Type":"ContainerDied","Data":"4b9eed3589fecc7f662df7a66d0e352b44a7d956645dcd7cbabdd8c9630814b7"} Oct 02 16:43:35 crc kubenswrapper[4808]: I1002 16:43:35.861158 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b9eed3589fecc7f662df7a66d0e352b44a7d956645dcd7cbabdd8c9630814b7" Oct 02 16:43:35 crc kubenswrapper[4808]: I1002 16:43:35.861305 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 16:43:40 crc kubenswrapper[4808]: I1002 16:43:40.793949 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 16:43:44 crc kubenswrapper[4808]: I1002 16:43:44.954037 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r4vs4" Oct 02 16:43:45 crc kubenswrapper[4808]: E1002 16:43:45.596776 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 02 16:43:45 crc kubenswrapper[4808]: E1002 16:43:45.597028 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6tqpg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-zbj7g_openshift-marketplace(09726ef2-b31b-48cd-8577-7eb3ae6aaca6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 16:43:45 crc kubenswrapper[4808]: E1002 16:43:45.598299 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-zbj7g" podUID="09726ef2-b31b-48cd-8577-7eb3ae6aaca6" Oct 02 16:43:46 crc kubenswrapper[4808]: E1002 16:43:46.094969 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 02 16:43:46 crc kubenswrapper[4808]: E1002 16:43:46.095182 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z52f6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-kbn54_openshift-marketplace(9439cd36-0f01-460d-a440-ac61df1bbb6f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 16:43:46 crc kubenswrapper[4808]: E1002 16:43:46.100423 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-kbn54" podUID="9439cd36-0f01-460d-a440-ac61df1bbb6f" Oct 02 16:43:48 crc kubenswrapper[4808]: E1002 16:43:48.095861 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-zbj7g" podUID="09726ef2-b31b-48cd-8577-7eb3ae6aaca6" Oct 02 16:43:48 crc kubenswrapper[4808]: E1002 16:43:48.096017 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-kbn54" podUID="9439cd36-0f01-460d-a440-ac61df1bbb6f" Oct 02 16:43:48 crc kubenswrapper[4808]: E1002 16:43:48.233879 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 02 16:43:48 crc kubenswrapper[4808]: E1002 16:43:48.234170 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c9jmk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-lhm6x_openshift-marketplace(429c5d29-dd2e-4938-a8cb-08a19212a6fc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 16:43:48 crc kubenswrapper[4808]: E1002 16:43:48.235543 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-lhm6x" podUID="429c5d29-dd2e-4938-a8cb-08a19212a6fc" Oct 02 16:43:48 crc kubenswrapper[4808]: E1002 16:43:48.584543 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 02 16:43:48 crc kubenswrapper[4808]: E1002 16:43:48.584793 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4pn24,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-nzsf5_openshift-marketplace(6c0f4487-6602-4c9d-b414-229588b2e797): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 16:43:48 crc kubenswrapper[4808]: E1002 16:43:48.586300 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-nzsf5" podUID="6c0f4487-6602-4c9d-b414-229588b2e797" Oct 02 16:43:48 crc kubenswrapper[4808]: I1002 16:43:48.740311 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 16:43:48 crc kubenswrapper[4808]: I1002 16:43:48.740407 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 16:43:52 crc kubenswrapper[4808]: E1002 16:43:52.656794 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-lhm6x" podUID="429c5d29-dd2e-4938-a8cb-08a19212a6fc" Oct 02 16:43:52 crc kubenswrapper[4808]: E1002 16:43:52.656883 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-nzsf5" podUID="6c0f4487-6602-4c9d-b414-229588b2e797" Oct 02 16:43:52 crc kubenswrapper[4808]: E1002 16:43:52.795990 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1: Get \"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1\": context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 16:43:52 crc kubenswrapper[4808]: E1002 16:43:52.796854 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-spv6h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-22l54_openshift-marketplace(25e17c62-b490-4c5d-8411-ef484d9039fc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1: Get \"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1\": context canceled" logger="UnhandledError" Oct 02 16:43:52 crc kubenswrapper[4808]: E1002 16:43:52.798560 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1: Get \\\"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1\\\": context canceled\"" pod="openshift-marketplace/redhat-marketplace-22l54" podUID="25e17c62-b490-4c5d-8411-ef484d9039fc" Oct 02 16:43:52 crc kubenswrapper[4808]: E1002 16:43:52.820103 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1: Get \"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1\": context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 16:43:52 crc kubenswrapper[4808]: E1002 16:43:52.820280 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bsjsh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-zhxtz_openshift-marketplace(fa0909cd-83b3-4127-abc6-2540f61e0b2b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1: Get \"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1\": context canceled" logger="UnhandledError" Oct 02 16:43:52 crc kubenswrapper[4808]: E1002 16:43:52.821466 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1: Get \\\"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1\\\": context canceled\"" pod="openshift-marketplace/redhat-marketplace-zhxtz" podUID="fa0909cd-83b3-4127-abc6-2540f61e0b2b" Oct 02 16:43:52 crc kubenswrapper[4808]: I1002 16:43:52.892630 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9vv2b"] Oct 02 16:43:52 crc kubenswrapper[4808]: I1002 16:43:52.973295 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" event={"ID":"71aafdd1-5917-4764-9503-4c266977f251","Type":"ContainerStarted","Data":"4387e5f3b5a1f67e79bf621dffeacc50959b134dd741e2a4102f50d751c048b7"} Oct 02 16:43:52 crc kubenswrapper[4808]: E1002 16:43:52.975173 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-22l54" podUID="25e17c62-b490-4c5d-8411-ef484d9039fc" Oct 02 16:43:52 crc kubenswrapper[4808]: E1002 16:43:52.976737 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-zhxtz" podUID="fa0909cd-83b3-4127-abc6-2540f61e0b2b" Oct 02 16:43:53 crc kubenswrapper[4808]: I1002 16:43:53.980685 4808 generic.go:334] "Generic (PLEG): container finished" podID="21467289-5a33-43ed-ac9b-425774019de0" containerID="6398818fd73d7f37e43e21bde66a7e1f69fc9b85f4cec1ab92b1526af73d4fb8" exitCode=0 Oct 02 16:43:53 crc kubenswrapper[4808]: I1002 16:43:53.980780 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dz6nh" event={"ID":"21467289-5a33-43ed-ac9b-425774019de0","Type":"ContainerDied","Data":"6398818fd73d7f37e43e21bde66a7e1f69fc9b85f4cec1ab92b1526af73d4fb8"} Oct 02 16:43:53 crc kubenswrapper[4808]: I1002 16:43:53.983187 4808 generic.go:334] "Generic (PLEG): container finished" podID="d621b4b0-80ad-489e-8f63-76918632c5ef" containerID="f548b392337ba3330769b010925bf36545282517b6475c48c3dcd0021fc67447" exitCode=0 Oct 02 16:43:53 crc kubenswrapper[4808]: I1002 16:43:53.983255 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwrlz" event={"ID":"d621b4b0-80ad-489e-8f63-76918632c5ef","Type":"ContainerDied","Data":"f548b392337ba3330769b010925bf36545282517b6475c48c3dcd0021fc67447"} Oct 02 16:43:53 crc kubenswrapper[4808]: I1002 16:43:53.987640 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" event={"ID":"71aafdd1-5917-4764-9503-4c266977f251","Type":"ContainerStarted","Data":"acc25e658141f487455f0b3f81462b05c303df4035f8ec9898137ee3282935fc"} Oct 02 16:43:53 crc kubenswrapper[4808]: I1002 16:43:53.988507 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:43:54 crc kubenswrapper[4808]: I1002 16:43:54.064459 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" podStartSLOduration=180.064436304 podStartE2EDuration="3m0.064436304s" podCreationTimestamp="2025-10-02 16:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:43:54.063467865 +0000 UTC m=+201.388996875" watchObservedRunningTime="2025-10-02 16:43:54.064436304 +0000 UTC m=+201.389965334" Oct 02 16:43:54 crc kubenswrapper[4808]: I1002 16:43:54.994915 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwrlz" event={"ID":"d621b4b0-80ad-489e-8f63-76918632c5ef","Type":"ContainerStarted","Data":"6fb37feb3ce1f537149b3c91476c54c6468369b1f02ec8c8e70e370f6cbb6979"} Oct 02 16:43:55 crc kubenswrapper[4808]: I1002 16:43:55.017267 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wwrlz" podStartSLOduration=5.210346905 podStartE2EDuration="42.017216979s" podCreationTimestamp="2025-10-02 16:43:13 +0000 UTC" firstStartedPulling="2025-10-02 16:43:17.720960345 +0000 UTC m=+165.046489355" lastFinishedPulling="2025-10-02 16:43:54.527830429 +0000 UTC m=+201.853359429" observedRunningTime="2025-10-02 16:43:55.012620092 +0000 UTC m=+202.338149152" watchObservedRunningTime="2025-10-02 16:43:55.017216979 +0000 UTC m=+202.342745999" Oct 02 16:43:56 crc kubenswrapper[4808]: I1002 16:43:56.006379 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dz6nh" event={"ID":"21467289-5a33-43ed-ac9b-425774019de0","Type":"ContainerStarted","Data":"b3043c438e926bf5a56d3705fe133d1d20c3d23a69a9b5ef67478bc06c8fef65"} Oct 02 16:43:56 crc kubenswrapper[4808]: I1002 16:43:56.028409 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dz6nh" podStartSLOduration=3.830626335 podStartE2EDuration="43.028383351s" podCreationTimestamp="2025-10-02 16:43:13 +0000 UTC" firstStartedPulling="2025-10-02 16:43:15.609649691 +0000 UTC m=+162.935178681" lastFinishedPulling="2025-10-02 16:43:54.807406697 +0000 UTC m=+202.132935697" observedRunningTime="2025-10-02 16:43:56.0243461 +0000 UTC m=+203.349875120" watchObservedRunningTime="2025-10-02 16:43:56.028383351 +0000 UTC m=+203.353912351" Oct 02 16:44:01 crc kubenswrapper[4808]: I1002 16:44:01.056270 4808 generic.go:334] "Generic (PLEG): container finished" podID="09726ef2-b31b-48cd-8577-7eb3ae6aaca6" containerID="1a997847ea15085c0d916c13b2d01c49a434825cf412ae59dda6b36649451caf" exitCode=0 Oct 02 16:44:01 crc kubenswrapper[4808]: I1002 16:44:01.056352 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zbj7g" event={"ID":"09726ef2-b31b-48cd-8577-7eb3ae6aaca6","Type":"ContainerDied","Data":"1a997847ea15085c0d916c13b2d01c49a434825cf412ae59dda6b36649451caf"} Oct 02 16:44:02 crc kubenswrapper[4808]: I1002 16:44:02.068503 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zbj7g" event={"ID":"09726ef2-b31b-48cd-8577-7eb3ae6aaca6","Type":"ContainerStarted","Data":"8a2b033564cdebddca3b9be513005c26512c68f558777c706f6a92aa4fd4be00"} Oct 02 16:44:02 crc kubenswrapper[4808]: I1002 16:44:02.097004 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zbj7g" podStartSLOduration=2.792606994 podStartE2EDuration="52.096984475s" podCreationTimestamp="2025-10-02 16:43:10 +0000 UTC" firstStartedPulling="2025-10-02 16:43:12.433612782 +0000 UTC m=+159.759141772" lastFinishedPulling="2025-10-02 16:44:01.737990253 +0000 UTC m=+209.063519253" observedRunningTime="2025-10-02 16:44:02.094004926 +0000 UTC m=+209.419533926" watchObservedRunningTime="2025-10-02 16:44:02.096984475 +0000 UTC m=+209.422513465" Oct 02 16:44:03 crc kubenswrapper[4808]: I1002 16:44:03.647550 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dz6nh" Oct 02 16:44:03 crc kubenswrapper[4808]: I1002 16:44:03.647999 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dz6nh" Oct 02 16:44:03 crc kubenswrapper[4808]: I1002 16:44:03.770741 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dz6nh" Oct 02 16:44:03 crc kubenswrapper[4808]: I1002 16:44:03.930244 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wwrlz" Oct 02 16:44:03 crc kubenswrapper[4808]: I1002 16:44:03.930611 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wwrlz" Oct 02 16:44:03 crc kubenswrapper[4808]: I1002 16:44:03.979422 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wwrlz" Oct 02 16:44:04 crc kubenswrapper[4808]: I1002 16:44:04.084219 4808 generic.go:334] "Generic (PLEG): container finished" podID="9439cd36-0f01-460d-a440-ac61df1bbb6f" containerID="ffdc7f04348f275f8bd31feade2bc2d06e466cf112876f71b01a69d1a215839a" exitCode=0 Oct 02 16:44:04 crc kubenswrapper[4808]: I1002 16:44:04.084374 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbn54" event={"ID":"9439cd36-0f01-460d-a440-ac61df1bbb6f","Type":"ContainerDied","Data":"ffdc7f04348f275f8bd31feade2bc2d06e466cf112876f71b01a69d1a215839a"} Oct 02 16:44:04 crc kubenswrapper[4808]: I1002 16:44:04.144665 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wwrlz" Oct 02 16:44:04 crc kubenswrapper[4808]: I1002 16:44:04.147046 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dz6nh" Oct 02 16:44:05 crc kubenswrapper[4808]: I1002 16:44:05.092294 4808 generic.go:334] "Generic (PLEG): container finished" podID="6c0f4487-6602-4c9d-b414-229588b2e797" containerID="464c5a518d670cf7c0a3bb5d7ede85bf8f86a8fa1b72e109e2a3d000b5f4f028" exitCode=0 Oct 02 16:44:05 crc kubenswrapper[4808]: I1002 16:44:05.093313 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nzsf5" event={"ID":"6c0f4487-6602-4c9d-b414-229588b2e797","Type":"ContainerDied","Data":"464c5a518d670cf7c0a3bb5d7ede85bf8f86a8fa1b72e109e2a3d000b5f4f028"} Oct 02 16:44:06 crc kubenswrapper[4808]: I1002 16:44:06.103827 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbn54" event={"ID":"9439cd36-0f01-460d-a440-ac61df1bbb6f","Type":"ContainerStarted","Data":"fe60ad368d70a235e0c0c5691bc96c25cb603e6d8be26c2a7ea86300c36cda9b"} Oct 02 16:44:06 crc kubenswrapper[4808]: I1002 16:44:06.113581 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wwrlz"] Oct 02 16:44:06 crc kubenswrapper[4808]: I1002 16:44:06.128316 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kbn54" podStartSLOduration=3.356056534 podStartE2EDuration="56.12828825s" podCreationTimestamp="2025-10-02 16:43:10 +0000 UTC" firstStartedPulling="2025-10-02 16:43:12.43275823 +0000 UTC m=+159.758287230" lastFinishedPulling="2025-10-02 16:44:05.204989946 +0000 UTC m=+212.530518946" observedRunningTime="2025-10-02 16:44:06.124877378 +0000 UTC m=+213.450406378" watchObservedRunningTime="2025-10-02 16:44:06.12828825 +0000 UTC m=+213.453817250" Oct 02 16:44:07 crc kubenswrapper[4808]: I1002 16:44:07.129673 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wwrlz" podUID="d621b4b0-80ad-489e-8f63-76918632c5ef" containerName="registry-server" containerID="cri-o://6fb37feb3ce1f537149b3c91476c54c6468369b1f02ec8c8e70e370f6cbb6979" gracePeriod=2 Oct 02 16:44:07 crc kubenswrapper[4808]: I1002 16:44:07.130291 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nzsf5" event={"ID":"6c0f4487-6602-4c9d-b414-229588b2e797","Type":"ContainerStarted","Data":"b9a880c7820137fd91e6b6a3bd59e14b8a4f4125675e748789d074caaa3a76c3"} Oct 02 16:44:07 crc kubenswrapper[4808]: I1002 16:44:07.157804 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nzsf5" podStartSLOduration=4.062964823 podStartE2EDuration="57.157782609s" podCreationTimestamp="2025-10-02 16:43:10 +0000 UTC" firstStartedPulling="2025-10-02 16:43:12.433829557 +0000 UTC m=+159.759358557" lastFinishedPulling="2025-10-02 16:44:05.528647343 +0000 UTC m=+212.854176343" observedRunningTime="2025-10-02 16:44:07.156564573 +0000 UTC m=+214.482093573" watchObservedRunningTime="2025-10-02 16:44:07.157782609 +0000 UTC m=+214.483311609" Oct 02 16:44:08 crc kubenswrapper[4808]: I1002 16:44:08.146486 4808 generic.go:334] "Generic (PLEG): container finished" podID="d621b4b0-80ad-489e-8f63-76918632c5ef" containerID="6fb37feb3ce1f537149b3c91476c54c6468369b1f02ec8c8e70e370f6cbb6979" exitCode=0 Oct 02 16:44:08 crc kubenswrapper[4808]: I1002 16:44:08.147313 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwrlz" event={"ID":"d621b4b0-80ad-489e-8f63-76918632c5ef","Type":"ContainerDied","Data":"6fb37feb3ce1f537149b3c91476c54c6468369b1f02ec8c8e70e370f6cbb6979"} Oct 02 16:44:09 crc kubenswrapper[4808]: I1002 16:44:09.242831 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:44:10 crc kubenswrapper[4808]: I1002 16:44:10.582905 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zbj7g" Oct 02 16:44:10 crc kubenswrapper[4808]: I1002 16:44:10.583541 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zbj7g" Oct 02 16:44:10 crc kubenswrapper[4808]: I1002 16:44:10.631342 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zbj7g" Oct 02 16:44:10 crc kubenswrapper[4808]: I1002 16:44:10.973828 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kbn54" Oct 02 16:44:10 crc kubenswrapper[4808]: I1002 16:44:10.973930 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kbn54" Oct 02 16:44:11 crc kubenswrapper[4808]: I1002 16:44:11.024435 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kbn54" Oct 02 16:44:11 crc kubenswrapper[4808]: I1002 16:44:11.210521 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kbn54" Oct 02 16:44:11 crc kubenswrapper[4808]: I1002 16:44:11.226414 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zbj7g" Oct 02 16:44:11 crc kubenswrapper[4808]: I1002 16:44:11.293421 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nzsf5" Oct 02 16:44:11 crc kubenswrapper[4808]: I1002 16:44:11.293576 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nzsf5" Oct 02 16:44:11 crc kubenswrapper[4808]: I1002 16:44:11.353419 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nzsf5" Oct 02 16:44:12 crc kubenswrapper[4808]: I1002 16:44:12.220051 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nzsf5" Oct 02 16:44:13 crc kubenswrapper[4808]: I1002 16:44:13.113872 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kbn54"] Oct 02 16:44:13 crc kubenswrapper[4808]: I1002 16:44:13.177081 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kbn54" podUID="9439cd36-0f01-460d-a440-ac61df1bbb6f" containerName="registry-server" containerID="cri-o://fe60ad368d70a235e0c0c5691bc96c25cb603e6d8be26c2a7ea86300c36cda9b" gracePeriod=2 Oct 02 16:44:13 crc kubenswrapper[4808]: E1002 16:44:13.930795 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6fb37feb3ce1f537149b3c91476c54c6468369b1f02ec8c8e70e370f6cbb6979 is running failed: container process not found" containerID="6fb37feb3ce1f537149b3c91476c54c6468369b1f02ec8c8e70e370f6cbb6979" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 16:44:13 crc kubenswrapper[4808]: E1002 16:44:13.932700 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6fb37feb3ce1f537149b3c91476c54c6468369b1f02ec8c8e70e370f6cbb6979 is running failed: container process not found" containerID="6fb37feb3ce1f537149b3c91476c54c6468369b1f02ec8c8e70e370f6cbb6979" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 16:44:13 crc kubenswrapper[4808]: E1002 16:44:13.933054 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6fb37feb3ce1f537149b3c91476c54c6468369b1f02ec8c8e70e370f6cbb6979 is running failed: container process not found" containerID="6fb37feb3ce1f537149b3c91476c54c6468369b1f02ec8c8e70e370f6cbb6979" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 16:44:13 crc kubenswrapper[4808]: E1002 16:44:13.933153 4808 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6fb37feb3ce1f537149b3c91476c54c6468369b1f02ec8c8e70e370f6cbb6979 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-wwrlz" podUID="d621b4b0-80ad-489e-8f63-76918632c5ef" containerName="registry-server" Oct 02 16:44:14 crc kubenswrapper[4808]: I1002 16:44:14.191693 4808 generic.go:334] "Generic (PLEG): container finished" podID="9439cd36-0f01-460d-a440-ac61df1bbb6f" containerID="fe60ad368d70a235e0c0c5691bc96c25cb603e6d8be26c2a7ea86300c36cda9b" exitCode=0 Oct 02 16:44:14 crc kubenswrapper[4808]: I1002 16:44:14.191744 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbn54" event={"ID":"9439cd36-0f01-460d-a440-ac61df1bbb6f","Type":"ContainerDied","Data":"fe60ad368d70a235e0c0c5691bc96c25cb603e6d8be26c2a7ea86300c36cda9b"} Oct 02 16:44:15 crc kubenswrapper[4808]: I1002 16:44:15.594382 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wwrlz" Oct 02 16:44:15 crc kubenswrapper[4808]: I1002 16:44:15.677759 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d621b4b0-80ad-489e-8f63-76918632c5ef-utilities\") pod \"d621b4b0-80ad-489e-8f63-76918632c5ef\" (UID: \"d621b4b0-80ad-489e-8f63-76918632c5ef\") " Oct 02 16:44:15 crc kubenswrapper[4808]: I1002 16:44:15.678224 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsszr\" (UniqueName: \"kubernetes.io/projected/d621b4b0-80ad-489e-8f63-76918632c5ef-kube-api-access-vsszr\") pod \"d621b4b0-80ad-489e-8f63-76918632c5ef\" (UID: \"d621b4b0-80ad-489e-8f63-76918632c5ef\") " Oct 02 16:44:15 crc kubenswrapper[4808]: I1002 16:44:15.678343 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d621b4b0-80ad-489e-8f63-76918632c5ef-catalog-content\") pod \"d621b4b0-80ad-489e-8f63-76918632c5ef\" (UID: \"d621b4b0-80ad-489e-8f63-76918632c5ef\") " Oct 02 16:44:15 crc kubenswrapper[4808]: I1002 16:44:15.679907 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d621b4b0-80ad-489e-8f63-76918632c5ef-utilities" (OuterVolumeSpecName: "utilities") pod "d621b4b0-80ad-489e-8f63-76918632c5ef" (UID: "d621b4b0-80ad-489e-8f63-76918632c5ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:44:15 crc kubenswrapper[4808]: I1002 16:44:15.697049 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d621b4b0-80ad-489e-8f63-76918632c5ef-kube-api-access-vsszr" (OuterVolumeSpecName: "kube-api-access-vsszr") pod "d621b4b0-80ad-489e-8f63-76918632c5ef" (UID: "d621b4b0-80ad-489e-8f63-76918632c5ef"). InnerVolumeSpecName "kube-api-access-vsszr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:44:15 crc kubenswrapper[4808]: I1002 16:44:15.780311 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsszr\" (UniqueName: \"kubernetes.io/projected/d621b4b0-80ad-489e-8f63-76918632c5ef-kube-api-access-vsszr\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:15 crc kubenswrapper[4808]: I1002 16:44:15.780351 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d621b4b0-80ad-489e-8f63-76918632c5ef-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:15 crc kubenswrapper[4808]: I1002 16:44:15.807926 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d621b4b0-80ad-489e-8f63-76918632c5ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d621b4b0-80ad-489e-8f63-76918632c5ef" (UID: "d621b4b0-80ad-489e-8f63-76918632c5ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:44:15 crc kubenswrapper[4808]: I1002 16:44:15.869180 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kbn54" Oct 02 16:44:15 crc kubenswrapper[4808]: I1002 16:44:15.881813 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d621b4b0-80ad-489e-8f63-76918632c5ef-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:15 crc kubenswrapper[4808]: I1002 16:44:15.982872 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z52f6\" (UniqueName: \"kubernetes.io/projected/9439cd36-0f01-460d-a440-ac61df1bbb6f-kube-api-access-z52f6\") pod \"9439cd36-0f01-460d-a440-ac61df1bbb6f\" (UID: \"9439cd36-0f01-460d-a440-ac61df1bbb6f\") " Oct 02 16:44:15 crc kubenswrapper[4808]: I1002 16:44:15.983034 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9439cd36-0f01-460d-a440-ac61df1bbb6f-catalog-content\") pod \"9439cd36-0f01-460d-a440-ac61df1bbb6f\" (UID: \"9439cd36-0f01-460d-a440-ac61df1bbb6f\") " Oct 02 16:44:15 crc kubenswrapper[4808]: I1002 16:44:15.983192 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9439cd36-0f01-460d-a440-ac61df1bbb6f-utilities\") pod \"9439cd36-0f01-460d-a440-ac61df1bbb6f\" (UID: \"9439cd36-0f01-460d-a440-ac61df1bbb6f\") " Oct 02 16:44:15 crc kubenswrapper[4808]: I1002 16:44:15.984078 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9439cd36-0f01-460d-a440-ac61df1bbb6f-utilities" (OuterVolumeSpecName: "utilities") pod "9439cd36-0f01-460d-a440-ac61df1bbb6f" (UID: "9439cd36-0f01-460d-a440-ac61df1bbb6f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:44:15 crc kubenswrapper[4808]: I1002 16:44:15.984679 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9439cd36-0f01-460d-a440-ac61df1bbb6f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:15 crc kubenswrapper[4808]: I1002 16:44:15.987082 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9439cd36-0f01-460d-a440-ac61df1bbb6f-kube-api-access-z52f6" (OuterVolumeSpecName: "kube-api-access-z52f6") pod "9439cd36-0f01-460d-a440-ac61df1bbb6f" (UID: "9439cd36-0f01-460d-a440-ac61df1bbb6f"). InnerVolumeSpecName "kube-api-access-z52f6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:44:16 crc kubenswrapper[4808]: I1002 16:44:16.090370 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z52f6\" (UniqueName: \"kubernetes.io/projected/9439cd36-0f01-460d-a440-ac61df1bbb6f-kube-api-access-z52f6\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:16 crc kubenswrapper[4808]: I1002 16:44:16.209349 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbn54" event={"ID":"9439cd36-0f01-460d-a440-ac61df1bbb6f","Type":"ContainerDied","Data":"3a259a0da166b9b5d132a46230967f9b403400de441f9134b4919047e000ef14"} Oct 02 16:44:16 crc kubenswrapper[4808]: I1002 16:44:16.209440 4808 scope.go:117] "RemoveContainer" containerID="fe60ad368d70a235e0c0c5691bc96c25cb603e6d8be26c2a7ea86300c36cda9b" Oct 02 16:44:16 crc kubenswrapper[4808]: I1002 16:44:16.209614 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kbn54" Oct 02 16:44:16 crc kubenswrapper[4808]: I1002 16:44:16.216730 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwrlz" event={"ID":"d621b4b0-80ad-489e-8f63-76918632c5ef","Type":"ContainerDied","Data":"bc11c5d52a0b7b4708a38ac1819b3803904741fa77edb1af52b1e5085ee102a0"} Oct 02 16:44:16 crc kubenswrapper[4808]: I1002 16:44:16.216835 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wwrlz" Oct 02 16:44:16 crc kubenswrapper[4808]: I1002 16:44:16.239030 4808 scope.go:117] "RemoveContainer" containerID="ffdc7f04348f275f8bd31feade2bc2d06e466cf112876f71b01a69d1a215839a" Oct 02 16:44:16 crc kubenswrapper[4808]: I1002 16:44:16.270589 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wwrlz"] Oct 02 16:44:16 crc kubenswrapper[4808]: I1002 16:44:16.274280 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wwrlz"] Oct 02 16:44:16 crc kubenswrapper[4808]: I1002 16:44:16.278483 4808 scope.go:117] "RemoveContainer" containerID="eb1d38a79c5806e30077f28e60f721d0c564bcf337b09e3b5f1b4e267d379b20" Oct 02 16:44:16 crc kubenswrapper[4808]: I1002 16:44:16.310969 4808 scope.go:117] "RemoveContainer" containerID="6fb37feb3ce1f537149b3c91476c54c6468369b1f02ec8c8e70e370f6cbb6979" Oct 02 16:44:16 crc kubenswrapper[4808]: I1002 16:44:16.324926 4808 scope.go:117] "RemoveContainer" containerID="f548b392337ba3330769b010925bf36545282517b6475c48c3dcd0021fc67447" Oct 02 16:44:16 crc kubenswrapper[4808]: I1002 16:44:16.344109 4808 scope.go:117] "RemoveContainer" containerID="3c6e9eb437a4c7e14d2595688255f944212621b20ad28d92717125e9dbde0170" Oct 02 16:44:16 crc kubenswrapper[4808]: I1002 16:44:16.445276 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9439cd36-0f01-460d-a440-ac61df1bbb6f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9439cd36-0f01-460d-a440-ac61df1bbb6f" (UID: "9439cd36-0f01-460d-a440-ac61df1bbb6f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:44:16 crc kubenswrapper[4808]: I1002 16:44:16.496980 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9439cd36-0f01-460d-a440-ac61df1bbb6f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:16 crc kubenswrapper[4808]: I1002 16:44:16.559593 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kbn54"] Oct 02 16:44:16 crc kubenswrapper[4808]: I1002 16:44:16.568180 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kbn54"] Oct 02 16:44:17 crc kubenswrapper[4808]: I1002 16:44:17.412804 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9439cd36-0f01-460d-a440-ac61df1bbb6f" path="/var/lib/kubelet/pods/9439cd36-0f01-460d-a440-ac61df1bbb6f/volumes" Oct 02 16:44:17 crc kubenswrapper[4808]: I1002 16:44:17.414929 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d621b4b0-80ad-489e-8f63-76918632c5ef" path="/var/lib/kubelet/pods/d621b4b0-80ad-489e-8f63-76918632c5ef/volumes" Oct 02 16:44:18 crc kubenswrapper[4808]: I1002 16:44:18.740175 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 16:44:18 crc kubenswrapper[4808]: I1002 16:44:18.740325 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 16:44:18 crc kubenswrapper[4808]: I1002 16:44:18.740428 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 16:44:18 crc kubenswrapper[4808]: I1002 16:44:18.741462 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba"} pod="openshift-machine-config-operator/machine-config-daemon-7z66r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 16:44:18 crc kubenswrapper[4808]: I1002 16:44:18.741661 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" containerID="cri-o://4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba" gracePeriod=600 Oct 02 16:44:19 crc kubenswrapper[4808]: I1002 16:44:19.249540 4808 generic.go:334] "Generic (PLEG): container finished" podID="f83ce425-101d-4489-94a4-5c256eb29328" containerID="4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba" exitCode=0 Oct 02 16:44:19 crc kubenswrapper[4808]: I1002 16:44:19.250111 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" event={"ID":"f83ce425-101d-4489-94a4-5c256eb29328","Type":"ContainerDied","Data":"4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba"} Oct 02 16:44:20 crc kubenswrapper[4808]: I1002 16:44:20.260265 4808 generic.go:334] "Generic (PLEG): container finished" podID="25e17c62-b490-4c5d-8411-ef484d9039fc" containerID="46b76faa1521f323708fc45e8259f1f27e164df4d1e58cf3003c98a436242f5a" exitCode=0 Oct 02 16:44:20 crc kubenswrapper[4808]: I1002 16:44:20.260376 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-22l54" event={"ID":"25e17c62-b490-4c5d-8411-ef484d9039fc","Type":"ContainerDied","Data":"46b76faa1521f323708fc45e8259f1f27e164df4d1e58cf3003c98a436242f5a"} Oct 02 16:44:20 crc kubenswrapper[4808]: I1002 16:44:20.266422 4808 generic.go:334] "Generic (PLEG): container finished" podID="fa0909cd-83b3-4127-abc6-2540f61e0b2b" containerID="2d0cef968392141b5021942cd9c8ea255ef06c79aa80c0ba9b3a1139fb2cc605" exitCode=0 Oct 02 16:44:20 crc kubenswrapper[4808]: I1002 16:44:20.266532 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhxtz" event={"ID":"fa0909cd-83b3-4127-abc6-2540f61e0b2b","Type":"ContainerDied","Data":"2d0cef968392141b5021942cd9c8ea255ef06c79aa80c0ba9b3a1139fb2cc605"} Oct 02 16:44:20 crc kubenswrapper[4808]: I1002 16:44:20.274036 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" event={"ID":"f83ce425-101d-4489-94a4-5c256eb29328","Type":"ContainerStarted","Data":"f801630a113a03b3e035f8940f01d306b49a7227476e62d1552f3229787c1924"} Oct 02 16:44:20 crc kubenswrapper[4808]: I1002 16:44:20.277531 4808 generic.go:334] "Generic (PLEG): container finished" podID="429c5d29-dd2e-4938-a8cb-08a19212a6fc" containerID="5cb8afd880e3342934161c25698f5dd3d49ba2bd436f55499d7c79c01bc39db3" exitCode=0 Oct 02 16:44:20 crc kubenswrapper[4808]: I1002 16:44:20.277582 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lhm6x" event={"ID":"429c5d29-dd2e-4938-a8cb-08a19212a6fc","Type":"ContainerDied","Data":"5cb8afd880e3342934161c25698f5dd3d49ba2bd436f55499d7c79c01bc39db3"} Oct 02 16:44:21 crc kubenswrapper[4808]: I1002 16:44:21.286030 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lhm6x" event={"ID":"429c5d29-dd2e-4938-a8cb-08a19212a6fc","Type":"ContainerStarted","Data":"19b9b8b128a6c631533de9c7c7a2d97bf36fd9b24bd472df0b664302df2a59d2"} Oct 02 16:44:21 crc kubenswrapper[4808]: I1002 16:44:21.290155 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-22l54" event={"ID":"25e17c62-b490-4c5d-8411-ef484d9039fc","Type":"ContainerStarted","Data":"e49b3aef7f2a99a3d073e8841813632b48c42a4ddd11d60234c5474e4701b155"} Oct 02 16:44:21 crc kubenswrapper[4808]: I1002 16:44:21.293781 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lhm6x" Oct 02 16:44:21 crc kubenswrapper[4808]: I1002 16:44:21.293837 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lhm6x" Oct 02 16:44:21 crc kubenswrapper[4808]: I1002 16:44:21.318925 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lhm6x" podStartSLOduration=2.869496436 podStartE2EDuration="1m11.318901696s" podCreationTimestamp="2025-10-02 16:43:10 +0000 UTC" firstStartedPulling="2025-10-02 16:43:12.447738977 +0000 UTC m=+159.773267977" lastFinishedPulling="2025-10-02 16:44:20.897144237 +0000 UTC m=+228.222673237" observedRunningTime="2025-10-02 16:44:21.314487954 +0000 UTC m=+228.640016954" watchObservedRunningTime="2025-10-02 16:44:21.318901696 +0000 UTC m=+228.644430696" Oct 02 16:44:21 crc kubenswrapper[4808]: I1002 16:44:21.342109 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-22l54" podStartSLOduration=2.893493153 podStartE2EDuration="1m9.342081189s" podCreationTimestamp="2025-10-02 16:43:12 +0000 UTC" firstStartedPulling="2025-10-02 16:43:14.579443104 +0000 UTC m=+161.904972104" lastFinishedPulling="2025-10-02 16:44:21.02803114 +0000 UTC m=+228.353560140" observedRunningTime="2025-10-02 16:44:21.335054259 +0000 UTC m=+228.660583269" watchObservedRunningTime="2025-10-02 16:44:21.342081189 +0000 UTC m=+228.667610199" Oct 02 16:44:22 crc kubenswrapper[4808]: I1002 16:44:22.300487 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhxtz" event={"ID":"fa0909cd-83b3-4127-abc6-2540f61e0b2b","Type":"ContainerStarted","Data":"e18ca276d6a428f6592fba1b8d263b238a8e25dfba04e6cde28de8eb28d55590"} Oct 02 16:44:22 crc kubenswrapper[4808]: I1002 16:44:22.335360 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zhxtz" podStartSLOduration=2.631303522 podStartE2EDuration="1m10.335325055s" podCreationTimestamp="2025-10-02 16:43:12 +0000 UTC" firstStartedPulling="2025-10-02 16:43:13.493794143 +0000 UTC m=+160.819323143" lastFinishedPulling="2025-10-02 16:44:21.197815676 +0000 UTC m=+228.523344676" observedRunningTime="2025-10-02 16:44:22.317250894 +0000 UTC m=+229.642779894" watchObservedRunningTime="2025-10-02 16:44:22.335325055 +0000 UTC m=+229.660854055" Oct 02 16:44:22 crc kubenswrapper[4808]: I1002 16:44:22.338081 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-lhm6x" podUID="429c5d29-dd2e-4938-a8cb-08a19212a6fc" containerName="registry-server" probeResult="failure" output=< Oct 02 16:44:22 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Oct 02 16:44:22 crc kubenswrapper[4808]: > Oct 02 16:44:22 crc kubenswrapper[4808]: I1002 16:44:22.518823 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zhxtz" Oct 02 16:44:22 crc kubenswrapper[4808]: I1002 16:44:22.519226 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zhxtz" Oct 02 16:44:22 crc kubenswrapper[4808]: I1002 16:44:22.570586 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zhxtz" Oct 02 16:44:22 crc kubenswrapper[4808]: I1002 16:44:22.958020 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-22l54" Oct 02 16:44:22 crc kubenswrapper[4808]: I1002 16:44:22.958066 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-22l54" Oct 02 16:44:23 crc kubenswrapper[4808]: I1002 16:44:23.005302 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-22l54" Oct 02 16:44:31 crc kubenswrapper[4808]: I1002 16:44:31.351722 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lhm6x" Oct 02 16:44:31 crc kubenswrapper[4808]: I1002 16:44:31.404763 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lhm6x" Oct 02 16:44:32 crc kubenswrapper[4808]: I1002 16:44:32.559771 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zhxtz" Oct 02 16:44:32 crc kubenswrapper[4808]: I1002 16:44:32.996522 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-22l54" Oct 02 16:44:33 crc kubenswrapper[4808]: I1002 16:44:33.715085 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lhm6x"] Oct 02 16:44:33 crc kubenswrapper[4808]: I1002 16:44:33.715902 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lhm6x" podUID="429c5d29-dd2e-4938-a8cb-08a19212a6fc" containerName="registry-server" containerID="cri-o://19b9b8b128a6c631533de9c7c7a2d97bf36fd9b24bd472df0b664302df2a59d2" gracePeriod=2 Oct 02 16:44:34 crc kubenswrapper[4808]: I1002 16:44:34.375398 4808 generic.go:334] "Generic (PLEG): container finished" podID="429c5d29-dd2e-4938-a8cb-08a19212a6fc" containerID="19b9b8b128a6c631533de9c7c7a2d97bf36fd9b24bd472df0b664302df2a59d2" exitCode=0 Oct 02 16:44:34 crc kubenswrapper[4808]: I1002 16:44:34.375455 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lhm6x" event={"ID":"429c5d29-dd2e-4938-a8cb-08a19212a6fc","Type":"ContainerDied","Data":"19b9b8b128a6c631533de9c7c7a2d97bf36fd9b24bd472df0b664302df2a59d2"} Oct 02 16:44:34 crc kubenswrapper[4808]: I1002 16:44:34.606618 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lhm6x" Oct 02 16:44:34 crc kubenswrapper[4808]: I1002 16:44:34.802264 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9jmk\" (UniqueName: \"kubernetes.io/projected/429c5d29-dd2e-4938-a8cb-08a19212a6fc-kube-api-access-c9jmk\") pod \"429c5d29-dd2e-4938-a8cb-08a19212a6fc\" (UID: \"429c5d29-dd2e-4938-a8cb-08a19212a6fc\") " Oct 02 16:44:34 crc kubenswrapper[4808]: I1002 16:44:34.802341 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/429c5d29-dd2e-4938-a8cb-08a19212a6fc-catalog-content\") pod \"429c5d29-dd2e-4938-a8cb-08a19212a6fc\" (UID: \"429c5d29-dd2e-4938-a8cb-08a19212a6fc\") " Oct 02 16:44:34 crc kubenswrapper[4808]: I1002 16:44:34.802360 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/429c5d29-dd2e-4938-a8cb-08a19212a6fc-utilities\") pod \"429c5d29-dd2e-4938-a8cb-08a19212a6fc\" (UID: \"429c5d29-dd2e-4938-a8cb-08a19212a6fc\") " Oct 02 16:44:34 crc kubenswrapper[4808]: I1002 16:44:34.803374 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/429c5d29-dd2e-4938-a8cb-08a19212a6fc-utilities" (OuterVolumeSpecName: "utilities") pod "429c5d29-dd2e-4938-a8cb-08a19212a6fc" (UID: "429c5d29-dd2e-4938-a8cb-08a19212a6fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:44:34 crc kubenswrapper[4808]: I1002 16:44:34.810547 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/429c5d29-dd2e-4938-a8cb-08a19212a6fc-kube-api-access-c9jmk" (OuterVolumeSpecName: "kube-api-access-c9jmk") pod "429c5d29-dd2e-4938-a8cb-08a19212a6fc" (UID: "429c5d29-dd2e-4938-a8cb-08a19212a6fc"). InnerVolumeSpecName "kube-api-access-c9jmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:44:34 crc kubenswrapper[4808]: I1002 16:44:34.857027 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/429c5d29-dd2e-4938-a8cb-08a19212a6fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "429c5d29-dd2e-4938-a8cb-08a19212a6fc" (UID: "429c5d29-dd2e-4938-a8cb-08a19212a6fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:44:34 crc kubenswrapper[4808]: I1002 16:44:34.904533 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/429c5d29-dd2e-4938-a8cb-08a19212a6fc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:34 crc kubenswrapper[4808]: I1002 16:44:34.904578 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/429c5d29-dd2e-4938-a8cb-08a19212a6fc-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:34 crc kubenswrapper[4808]: I1002 16:44:34.904588 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9jmk\" (UniqueName: \"kubernetes.io/projected/429c5d29-dd2e-4938-a8cb-08a19212a6fc-kube-api-access-c9jmk\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:35 crc kubenswrapper[4808]: I1002 16:44:35.385107 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lhm6x" event={"ID":"429c5d29-dd2e-4938-a8cb-08a19212a6fc","Type":"ContainerDied","Data":"83fbb3eae0ed5fdd3bcea86977934130f2ea839bfc689e1d0d1af90a7699815c"} Oct 02 16:44:35 crc kubenswrapper[4808]: I1002 16:44:35.385218 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lhm6x" Oct 02 16:44:35 crc kubenswrapper[4808]: I1002 16:44:35.385552 4808 scope.go:117] "RemoveContainer" containerID="19b9b8b128a6c631533de9c7c7a2d97bf36fd9b24bd472df0b664302df2a59d2" Oct 02 16:44:35 crc kubenswrapper[4808]: I1002 16:44:35.405333 4808 scope.go:117] "RemoveContainer" containerID="5cb8afd880e3342934161c25698f5dd3d49ba2bd436f55499d7c79c01bc39db3" Oct 02 16:44:35 crc kubenswrapper[4808]: I1002 16:44:35.428125 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lhm6x"] Oct 02 16:44:35 crc kubenswrapper[4808]: I1002 16:44:35.428198 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lhm6x"] Oct 02 16:44:35 crc kubenswrapper[4808]: I1002 16:44:35.430273 4808 scope.go:117] "RemoveContainer" containerID="2bba5502fe52cfc26051e1235f5122f585641a61caa7d7dc5d00ab6c3a59dba7" Oct 02 16:44:36 crc kubenswrapper[4808]: I1002 16:44:36.114604 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-22l54"] Oct 02 16:44:36 crc kubenswrapper[4808]: I1002 16:44:36.115786 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-22l54" podUID="25e17c62-b490-4c5d-8411-ef484d9039fc" containerName="registry-server" containerID="cri-o://e49b3aef7f2a99a3d073e8841813632b48c42a4ddd11d60234c5474e4701b155" gracePeriod=2 Oct 02 16:44:37 crc kubenswrapper[4808]: I1002 16:44:37.402914 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="429c5d29-dd2e-4938-a8cb-08a19212a6fc" path="/var/lib/kubelet/pods/429c5d29-dd2e-4938-a8cb-08a19212a6fc/volumes" Oct 02 16:44:39 crc kubenswrapper[4808]: I1002 16:44:39.417975 4808 generic.go:334] "Generic (PLEG): container finished" podID="25e17c62-b490-4c5d-8411-ef484d9039fc" containerID="e49b3aef7f2a99a3d073e8841813632b48c42a4ddd11d60234c5474e4701b155" exitCode=0 Oct 02 16:44:39 crc kubenswrapper[4808]: I1002 16:44:39.418090 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-22l54" event={"ID":"25e17c62-b490-4c5d-8411-ef484d9039fc","Type":"ContainerDied","Data":"e49b3aef7f2a99a3d073e8841813632b48c42a4ddd11d60234c5474e4701b155"} Oct 02 16:44:39 crc kubenswrapper[4808]: I1002 16:44:39.714493 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-22l54" Oct 02 16:44:39 crc kubenswrapper[4808]: I1002 16:44:39.889908 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spv6h\" (UniqueName: \"kubernetes.io/projected/25e17c62-b490-4c5d-8411-ef484d9039fc-kube-api-access-spv6h\") pod \"25e17c62-b490-4c5d-8411-ef484d9039fc\" (UID: \"25e17c62-b490-4c5d-8411-ef484d9039fc\") " Oct 02 16:44:39 crc kubenswrapper[4808]: I1002 16:44:39.890106 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25e17c62-b490-4c5d-8411-ef484d9039fc-catalog-content\") pod \"25e17c62-b490-4c5d-8411-ef484d9039fc\" (UID: \"25e17c62-b490-4c5d-8411-ef484d9039fc\") " Oct 02 16:44:39 crc kubenswrapper[4808]: I1002 16:44:39.890173 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25e17c62-b490-4c5d-8411-ef484d9039fc-utilities\") pod \"25e17c62-b490-4c5d-8411-ef484d9039fc\" (UID: \"25e17c62-b490-4c5d-8411-ef484d9039fc\") " Oct 02 16:44:39 crc kubenswrapper[4808]: I1002 16:44:39.891759 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25e17c62-b490-4c5d-8411-ef484d9039fc-utilities" (OuterVolumeSpecName: "utilities") pod "25e17c62-b490-4c5d-8411-ef484d9039fc" (UID: "25e17c62-b490-4c5d-8411-ef484d9039fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:44:39 crc kubenswrapper[4808]: I1002 16:44:39.901101 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e17c62-b490-4c5d-8411-ef484d9039fc-kube-api-access-spv6h" (OuterVolumeSpecName: "kube-api-access-spv6h") pod "25e17c62-b490-4c5d-8411-ef484d9039fc" (UID: "25e17c62-b490-4c5d-8411-ef484d9039fc"). InnerVolumeSpecName "kube-api-access-spv6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:44:39 crc kubenswrapper[4808]: I1002 16:44:39.910348 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25e17c62-b490-4c5d-8411-ef484d9039fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25e17c62-b490-4c5d-8411-ef484d9039fc" (UID: "25e17c62-b490-4c5d-8411-ef484d9039fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:44:39 crc kubenswrapper[4808]: I1002 16:44:39.992369 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spv6h\" (UniqueName: \"kubernetes.io/projected/25e17c62-b490-4c5d-8411-ef484d9039fc-kube-api-access-spv6h\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:39 crc kubenswrapper[4808]: I1002 16:44:39.992433 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25e17c62-b490-4c5d-8411-ef484d9039fc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:39 crc kubenswrapper[4808]: I1002 16:44:39.992453 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25e17c62-b490-4c5d-8411-ef484d9039fc-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:40 crc kubenswrapper[4808]: I1002 16:44:40.429111 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-22l54" event={"ID":"25e17c62-b490-4c5d-8411-ef484d9039fc","Type":"ContainerDied","Data":"39a7dff4a485fa17cfd97a19a5e6b645de548f1618b2b7f7926ebb7bcb90710a"} Oct 02 16:44:40 crc kubenswrapper[4808]: I1002 16:44:40.429669 4808 scope.go:117] "RemoveContainer" containerID="e49b3aef7f2a99a3d073e8841813632b48c42a4ddd11d60234c5474e4701b155" Oct 02 16:44:40 crc kubenswrapper[4808]: I1002 16:44:40.429265 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-22l54" Oct 02 16:44:40 crc kubenswrapper[4808]: I1002 16:44:40.455735 4808 scope.go:117] "RemoveContainer" containerID="46b76faa1521f323708fc45e8259f1f27e164df4d1e58cf3003c98a436242f5a" Oct 02 16:44:40 crc kubenswrapper[4808]: I1002 16:44:40.482821 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-22l54"] Oct 02 16:44:40 crc kubenswrapper[4808]: I1002 16:44:40.482884 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-22l54"] Oct 02 16:44:40 crc kubenswrapper[4808]: I1002 16:44:40.507600 4808 scope.go:117] "RemoveContainer" containerID="f47657a5c62497b5768bd4f0d8ea00c5770c779e63e445b7d4e40bd18bd8f291" Oct 02 16:44:41 crc kubenswrapper[4808]: I1002 16:44:41.413415 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e17c62-b490-4c5d-8411-ef484d9039fc" path="/var/lib/kubelet/pods/25e17c62-b490-4c5d-8411-ef484d9039fc/volumes" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.773044 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zbj7g"] Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.774498 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zbj7g" podUID="09726ef2-b31b-48cd-8577-7eb3ae6aaca6" containerName="registry-server" containerID="cri-o://8a2b033564cdebddca3b9be513005c26512c68f558777c706f6a92aa4fd4be00" gracePeriod=30 Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.794937 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nzsf5"] Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.795703 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nzsf5" podUID="6c0f4487-6602-4c9d-b414-229588b2e797" containerName="registry-server" containerID="cri-o://b9a880c7820137fd91e6b6a3bd59e14b8a4f4125675e748789d074caaa3a76c3" gracePeriod=30 Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.801992 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gpxn2"] Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.804938 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" podUID="02097ffc-aa46-438f-b11d-c98301f3e617" containerName="marketplace-operator" containerID="cri-o://c4917adf048a4ec961ca2f6c21b80a0dddab2f112a5bc63ba2a21834b63ce6cc" gracePeriod=30 Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.816890 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhxtz"] Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.817267 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zhxtz" podUID="fa0909cd-83b3-4127-abc6-2540f61e0b2b" containerName="registry-server" containerID="cri-o://e18ca276d6a428f6592fba1b8d263b238a8e25dfba04e6cde28de8eb28d55590" gracePeriod=30 Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.831868 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dz6nh"] Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.832167 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dz6nh" podUID="21467289-5a33-43ed-ac9b-425774019de0" containerName="registry-server" containerID="cri-o://b3043c438e926bf5a56d3705fe133d1d20c3d23a69a9b5ef67478bc06c8fef65" gracePeriod=30 Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839277 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bhdpb"] Oct 02 16:44:52 crc kubenswrapper[4808]: E1002 16:44:52.839552 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00ff555d-8628-48a3-b667-a2b5d1aeb249" containerName="pruner" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839568 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="00ff555d-8628-48a3-b667-a2b5d1aeb249" containerName="pruner" Oct 02 16:44:52 crc kubenswrapper[4808]: E1002 16:44:52.839578 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25e17c62-b490-4c5d-8411-ef484d9039fc" containerName="extract-utilities" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839586 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="25e17c62-b490-4c5d-8411-ef484d9039fc" containerName="extract-utilities" Oct 02 16:44:52 crc kubenswrapper[4808]: E1002 16:44:52.839595 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9439cd36-0f01-460d-a440-ac61df1bbb6f" containerName="extract-content" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839601 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9439cd36-0f01-460d-a440-ac61df1bbb6f" containerName="extract-content" Oct 02 16:44:52 crc kubenswrapper[4808]: E1002 16:44:52.839609 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25e17c62-b490-4c5d-8411-ef484d9039fc" containerName="registry-server" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839616 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="25e17c62-b490-4c5d-8411-ef484d9039fc" containerName="registry-server" Oct 02 16:44:52 crc kubenswrapper[4808]: E1002 16:44:52.839624 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="429c5d29-dd2e-4938-a8cb-08a19212a6fc" containerName="extract-utilities" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839631 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="429c5d29-dd2e-4938-a8cb-08a19212a6fc" containerName="extract-utilities" Oct 02 16:44:52 crc kubenswrapper[4808]: E1002 16:44:52.839639 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25e17c62-b490-4c5d-8411-ef484d9039fc" containerName="extract-content" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839648 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="25e17c62-b490-4c5d-8411-ef484d9039fc" containerName="extract-content" Oct 02 16:44:52 crc kubenswrapper[4808]: E1002 16:44:52.839658 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="429c5d29-dd2e-4938-a8cb-08a19212a6fc" containerName="extract-content" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839666 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="429c5d29-dd2e-4938-a8cb-08a19212a6fc" containerName="extract-content" Oct 02 16:44:52 crc kubenswrapper[4808]: E1002 16:44:52.839679 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9439cd36-0f01-460d-a440-ac61df1bbb6f" containerName="registry-server" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839685 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9439cd36-0f01-460d-a440-ac61df1bbb6f" containerName="registry-server" Oct 02 16:44:52 crc kubenswrapper[4808]: E1002 16:44:52.839693 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9439cd36-0f01-460d-a440-ac61df1bbb6f" containerName="extract-utilities" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839701 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9439cd36-0f01-460d-a440-ac61df1bbb6f" containerName="extract-utilities" Oct 02 16:44:52 crc kubenswrapper[4808]: E1002 16:44:52.839713 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="429c5d29-dd2e-4938-a8cb-08a19212a6fc" containerName="registry-server" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839721 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="429c5d29-dd2e-4938-a8cb-08a19212a6fc" containerName="registry-server" Oct 02 16:44:52 crc kubenswrapper[4808]: E1002 16:44:52.839733 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d621b4b0-80ad-489e-8f63-76918632c5ef" containerName="extract-utilities" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839742 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d621b4b0-80ad-489e-8f63-76918632c5ef" containerName="extract-utilities" Oct 02 16:44:52 crc kubenswrapper[4808]: E1002 16:44:52.839754 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d621b4b0-80ad-489e-8f63-76918632c5ef" containerName="extract-content" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839761 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d621b4b0-80ad-489e-8f63-76918632c5ef" containerName="extract-content" Oct 02 16:44:52 crc kubenswrapper[4808]: E1002 16:44:52.839773 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d621b4b0-80ad-489e-8f63-76918632c5ef" containerName="registry-server" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839780 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d621b4b0-80ad-489e-8f63-76918632c5ef" containerName="registry-server" Oct 02 16:44:52 crc kubenswrapper[4808]: E1002 16:44:52.839795 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6bc16bd-85c9-4ec5-b636-38b69f616aa8" containerName="pruner" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839803 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6bc16bd-85c9-4ec5-b636-38b69f616aa8" containerName="pruner" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839939 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="d621b4b0-80ad-489e-8f63-76918632c5ef" containerName="registry-server" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839951 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6bc16bd-85c9-4ec5-b636-38b69f616aa8" containerName="pruner" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839959 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="9439cd36-0f01-460d-a440-ac61df1bbb6f" containerName="registry-server" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839970 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="00ff555d-8628-48a3-b667-a2b5d1aeb249" containerName="pruner" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839978 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="429c5d29-dd2e-4938-a8cb-08a19212a6fc" containerName="registry-server" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.839986 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="25e17c62-b490-4c5d-8411-ef484d9039fc" containerName="registry-server" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.842716 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bhdpb" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.845459 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bhdpb"] Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.988504 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/037881d6-84ee-47ba-a0cf-f6ac68f7ef17-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bhdpb\" (UID: \"037881d6-84ee-47ba-a0cf-f6ac68f7ef17\") " pod="openshift-marketplace/marketplace-operator-79b997595-bhdpb" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.988590 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4fpr\" (UniqueName: \"kubernetes.io/projected/037881d6-84ee-47ba-a0cf-f6ac68f7ef17-kube-api-access-z4fpr\") pod \"marketplace-operator-79b997595-bhdpb\" (UID: \"037881d6-84ee-47ba-a0cf-f6ac68f7ef17\") " pod="openshift-marketplace/marketplace-operator-79b997595-bhdpb" Oct 02 16:44:52 crc kubenswrapper[4808]: I1002 16:44:52.988664 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/037881d6-84ee-47ba-a0cf-f6ac68f7ef17-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bhdpb\" (UID: \"037881d6-84ee-47ba-a0cf-f6ac68f7ef17\") " pod="openshift-marketplace/marketplace-operator-79b997595-bhdpb" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.097629 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/037881d6-84ee-47ba-a0cf-f6ac68f7ef17-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bhdpb\" (UID: \"037881d6-84ee-47ba-a0cf-f6ac68f7ef17\") " pod="openshift-marketplace/marketplace-operator-79b997595-bhdpb" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.098025 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/037881d6-84ee-47ba-a0cf-f6ac68f7ef17-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bhdpb\" (UID: \"037881d6-84ee-47ba-a0cf-f6ac68f7ef17\") " pod="openshift-marketplace/marketplace-operator-79b997595-bhdpb" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.098063 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4fpr\" (UniqueName: \"kubernetes.io/projected/037881d6-84ee-47ba-a0cf-f6ac68f7ef17-kube-api-access-z4fpr\") pod \"marketplace-operator-79b997595-bhdpb\" (UID: \"037881d6-84ee-47ba-a0cf-f6ac68f7ef17\") " pod="openshift-marketplace/marketplace-operator-79b997595-bhdpb" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.099589 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/037881d6-84ee-47ba-a0cf-f6ac68f7ef17-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bhdpb\" (UID: \"037881d6-84ee-47ba-a0cf-f6ac68f7ef17\") " pod="openshift-marketplace/marketplace-operator-79b997595-bhdpb" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.105363 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/037881d6-84ee-47ba-a0cf-f6ac68f7ef17-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bhdpb\" (UID: \"037881d6-84ee-47ba-a0cf-f6ac68f7ef17\") " pod="openshift-marketplace/marketplace-operator-79b997595-bhdpb" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.116744 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4fpr\" (UniqueName: \"kubernetes.io/projected/037881d6-84ee-47ba-a0cf-f6ac68f7ef17-kube-api-access-z4fpr\") pod \"marketplace-operator-79b997595-bhdpb\" (UID: \"037881d6-84ee-47ba-a0cf-f6ac68f7ef17\") " pod="openshift-marketplace/marketplace-operator-79b997595-bhdpb" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.283027 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bhdpb" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.287254 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zbj7g" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.297881 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nzsf5" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.299924 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.300542 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tqpg\" (UniqueName: \"kubernetes.io/projected/09726ef2-b31b-48cd-8577-7eb3ae6aaca6-kube-api-access-6tqpg\") pod \"09726ef2-b31b-48cd-8577-7eb3ae6aaca6\" (UID: \"09726ef2-b31b-48cd-8577-7eb3ae6aaca6\") " Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.300681 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09726ef2-b31b-48cd-8577-7eb3ae6aaca6-catalog-content\") pod \"09726ef2-b31b-48cd-8577-7eb3ae6aaca6\" (UID: \"09726ef2-b31b-48cd-8577-7eb3ae6aaca6\") " Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.300847 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09726ef2-b31b-48cd-8577-7eb3ae6aaca6-utilities\") pod \"09726ef2-b31b-48cd-8577-7eb3ae6aaca6\" (UID: \"09726ef2-b31b-48cd-8577-7eb3ae6aaca6\") " Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.302267 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09726ef2-b31b-48cd-8577-7eb3ae6aaca6-utilities" (OuterVolumeSpecName: "utilities") pod "09726ef2-b31b-48cd-8577-7eb3ae6aaca6" (UID: "09726ef2-b31b-48cd-8577-7eb3ae6aaca6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.304084 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09726ef2-b31b-48cd-8577-7eb3ae6aaca6-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.305261 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09726ef2-b31b-48cd-8577-7eb3ae6aaca6-kube-api-access-6tqpg" (OuterVolumeSpecName: "kube-api-access-6tqpg") pod "09726ef2-b31b-48cd-8577-7eb3ae6aaca6" (UID: "09726ef2-b31b-48cd-8577-7eb3ae6aaca6"). InnerVolumeSpecName "kube-api-access-6tqpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.337285 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zhxtz" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.341775 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dz6nh" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.383351 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09726ef2-b31b-48cd-8577-7eb3ae6aaca6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09726ef2-b31b-48cd-8577-7eb3ae6aaca6" (UID: "09726ef2-b31b-48cd-8577-7eb3ae6aaca6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.404657 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa0909cd-83b3-4127-abc6-2540f61e0b2b-catalog-content\") pod \"fa0909cd-83b3-4127-abc6-2540f61e0b2b\" (UID: \"fa0909cd-83b3-4127-abc6-2540f61e0b2b\") " Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.404739 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/02097ffc-aa46-438f-b11d-c98301f3e617-marketplace-trusted-ca\") pod \"02097ffc-aa46-438f-b11d-c98301f3e617\" (UID: \"02097ffc-aa46-438f-b11d-c98301f3e617\") " Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.404781 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc7rr\" (UniqueName: \"kubernetes.io/projected/02097ffc-aa46-438f-b11d-c98301f3e617-kube-api-access-rc7rr\") pod \"02097ffc-aa46-438f-b11d-c98301f3e617\" (UID: \"02097ffc-aa46-438f-b11d-c98301f3e617\") " Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.404850 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rr54f\" (UniqueName: \"kubernetes.io/projected/21467289-5a33-43ed-ac9b-425774019de0-kube-api-access-rr54f\") pod \"21467289-5a33-43ed-ac9b-425774019de0\" (UID: \"21467289-5a33-43ed-ac9b-425774019de0\") " Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.404895 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/02097ffc-aa46-438f-b11d-c98301f3e617-marketplace-operator-metrics\") pod \"02097ffc-aa46-438f-b11d-c98301f3e617\" (UID: \"02097ffc-aa46-438f-b11d-c98301f3e617\") " Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.404926 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c0f4487-6602-4c9d-b414-229588b2e797-catalog-content\") pod \"6c0f4487-6602-4c9d-b414-229588b2e797\" (UID: \"6c0f4487-6602-4c9d-b414-229588b2e797\") " Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.404953 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa0909cd-83b3-4127-abc6-2540f61e0b2b-utilities\") pod \"fa0909cd-83b3-4127-abc6-2540f61e0b2b\" (UID: \"fa0909cd-83b3-4127-abc6-2540f61e0b2b\") " Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.404977 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsjsh\" (UniqueName: \"kubernetes.io/projected/fa0909cd-83b3-4127-abc6-2540f61e0b2b-kube-api-access-bsjsh\") pod \"fa0909cd-83b3-4127-abc6-2540f61e0b2b\" (UID: \"fa0909cd-83b3-4127-abc6-2540f61e0b2b\") " Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.405017 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pn24\" (UniqueName: \"kubernetes.io/projected/6c0f4487-6602-4c9d-b414-229588b2e797-kube-api-access-4pn24\") pod \"6c0f4487-6602-4c9d-b414-229588b2e797\" (UID: \"6c0f4487-6602-4c9d-b414-229588b2e797\") " Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.405122 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21467289-5a33-43ed-ac9b-425774019de0-catalog-content\") pod \"21467289-5a33-43ed-ac9b-425774019de0\" (UID: \"21467289-5a33-43ed-ac9b-425774019de0\") " Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.405262 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21467289-5a33-43ed-ac9b-425774019de0-utilities\") pod \"21467289-5a33-43ed-ac9b-425774019de0\" (UID: \"21467289-5a33-43ed-ac9b-425774019de0\") " Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.405309 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c0f4487-6602-4c9d-b414-229588b2e797-utilities\") pod \"6c0f4487-6602-4c9d-b414-229588b2e797\" (UID: \"6c0f4487-6602-4c9d-b414-229588b2e797\") " Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.405629 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tqpg\" (UniqueName: \"kubernetes.io/projected/09726ef2-b31b-48cd-8577-7eb3ae6aaca6-kube-api-access-6tqpg\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.405657 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09726ef2-b31b-48cd-8577-7eb3ae6aaca6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.406604 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c0f4487-6602-4c9d-b414-229588b2e797-utilities" (OuterVolumeSpecName: "utilities") pod "6c0f4487-6602-4c9d-b414-229588b2e797" (UID: "6c0f4487-6602-4c9d-b414-229588b2e797"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.409249 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa0909cd-83b3-4127-abc6-2540f61e0b2b-kube-api-access-bsjsh" (OuterVolumeSpecName: "kube-api-access-bsjsh") pod "fa0909cd-83b3-4127-abc6-2540f61e0b2b" (UID: "fa0909cd-83b3-4127-abc6-2540f61e0b2b"). InnerVolumeSpecName "kube-api-access-bsjsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.413549 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21467289-5a33-43ed-ac9b-425774019de0-utilities" (OuterVolumeSpecName: "utilities") pod "21467289-5a33-43ed-ac9b-425774019de0" (UID: "21467289-5a33-43ed-ac9b-425774019de0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.413783 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02097ffc-aa46-438f-b11d-c98301f3e617-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "02097ffc-aa46-438f-b11d-c98301f3e617" (UID: "02097ffc-aa46-438f-b11d-c98301f3e617"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.414264 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02097ffc-aa46-438f-b11d-c98301f3e617-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "02097ffc-aa46-438f-b11d-c98301f3e617" (UID: "02097ffc-aa46-438f-b11d-c98301f3e617"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.416091 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa0909cd-83b3-4127-abc6-2540f61e0b2b-utilities" (OuterVolumeSpecName: "utilities") pod "fa0909cd-83b3-4127-abc6-2540f61e0b2b" (UID: "fa0909cd-83b3-4127-abc6-2540f61e0b2b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.423992 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21467289-5a33-43ed-ac9b-425774019de0-kube-api-access-rr54f" (OuterVolumeSpecName: "kube-api-access-rr54f") pod "21467289-5a33-43ed-ac9b-425774019de0" (UID: "21467289-5a33-43ed-ac9b-425774019de0"). InnerVolumeSpecName "kube-api-access-rr54f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.431156 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02097ffc-aa46-438f-b11d-c98301f3e617-kube-api-access-rc7rr" (OuterVolumeSpecName: "kube-api-access-rc7rr") pod "02097ffc-aa46-438f-b11d-c98301f3e617" (UID: "02097ffc-aa46-438f-b11d-c98301f3e617"). InnerVolumeSpecName "kube-api-access-rc7rr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.446120 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c0f4487-6602-4c9d-b414-229588b2e797-kube-api-access-4pn24" (OuterVolumeSpecName: "kube-api-access-4pn24") pod "6c0f4487-6602-4c9d-b414-229588b2e797" (UID: "6c0f4487-6602-4c9d-b414-229588b2e797"). InnerVolumeSpecName "kube-api-access-4pn24". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.471746 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa0909cd-83b3-4127-abc6-2540f61e0b2b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa0909cd-83b3-4127-abc6-2540f61e0b2b" (UID: "fa0909cd-83b3-4127-abc6-2540f61e0b2b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.482442 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c0f4487-6602-4c9d-b414-229588b2e797-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6c0f4487-6602-4c9d-b414-229588b2e797" (UID: "6c0f4487-6602-4c9d-b414-229588b2e797"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.508278 4808 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/02097ffc-aa46-438f-b11d-c98301f3e617-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.508639 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc7rr\" (UniqueName: \"kubernetes.io/projected/02097ffc-aa46-438f-b11d-c98301f3e617-kube-api-access-rc7rr\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.508723 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rr54f\" (UniqueName: \"kubernetes.io/projected/21467289-5a33-43ed-ac9b-425774019de0-kube-api-access-rr54f\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.508795 4808 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/02097ffc-aa46-438f-b11d-c98301f3e617-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.508874 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c0f4487-6602-4c9d-b414-229588b2e797-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.508999 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa0909cd-83b3-4127-abc6-2540f61e0b2b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.509087 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsjsh\" (UniqueName: \"kubernetes.io/projected/fa0909cd-83b3-4127-abc6-2540f61e0b2b-kube-api-access-bsjsh\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.509170 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pn24\" (UniqueName: \"kubernetes.io/projected/6c0f4487-6602-4c9d-b414-229588b2e797-kube-api-access-4pn24\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.509281 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21467289-5a33-43ed-ac9b-425774019de0-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.509369 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c0f4487-6602-4c9d-b414-229588b2e797-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.509445 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa0909cd-83b3-4127-abc6-2540f61e0b2b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.515908 4808 generic.go:334] "Generic (PLEG): container finished" podID="fa0909cd-83b3-4127-abc6-2540f61e0b2b" containerID="e18ca276d6a428f6592fba1b8d263b238a8e25dfba04e6cde28de8eb28d55590" exitCode=0 Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.516070 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zhxtz" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.516267 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhxtz" event={"ID":"fa0909cd-83b3-4127-abc6-2540f61e0b2b","Type":"ContainerDied","Data":"e18ca276d6a428f6592fba1b8d263b238a8e25dfba04e6cde28de8eb28d55590"} Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.516366 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhxtz" event={"ID":"fa0909cd-83b3-4127-abc6-2540f61e0b2b","Type":"ContainerDied","Data":"93b68e787fb922d50207f8ee2c5cac6ee72af1f3a9790354ed4cd09a42d0a668"} Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.516408 4808 scope.go:117] "RemoveContainer" containerID="e18ca276d6a428f6592fba1b8d263b238a8e25dfba04e6cde28de8eb28d55590" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.550472 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21467289-5a33-43ed-ac9b-425774019de0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "21467289-5a33-43ed-ac9b-425774019de0" (UID: "21467289-5a33-43ed-ac9b-425774019de0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.553755 4808 generic.go:334] "Generic (PLEG): container finished" podID="6c0f4487-6602-4c9d-b414-229588b2e797" containerID="b9a880c7820137fd91e6b6a3bd59e14b8a4f4125675e748789d074caaa3a76c3" exitCode=0 Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.553857 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nzsf5" event={"ID":"6c0f4487-6602-4c9d-b414-229588b2e797","Type":"ContainerDied","Data":"b9a880c7820137fd91e6b6a3bd59e14b8a4f4125675e748789d074caaa3a76c3"} Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.553892 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nzsf5" event={"ID":"6c0f4487-6602-4c9d-b414-229588b2e797","Type":"ContainerDied","Data":"9428346d2a142aa8412b3b30ff6bc3094afa01b306abc0f68c150d16d2af7329"} Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.553996 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nzsf5" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.560655 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bhdpb"] Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.566633 4808 scope.go:117] "RemoveContainer" containerID="2d0cef968392141b5021942cd9c8ea255ef06c79aa80c0ba9b3a1139fb2cc605" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.570745 4808 generic.go:334] "Generic (PLEG): container finished" podID="09726ef2-b31b-48cd-8577-7eb3ae6aaca6" containerID="8a2b033564cdebddca3b9be513005c26512c68f558777c706f6a92aa4fd4be00" exitCode=0 Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.570819 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zbj7g" event={"ID":"09726ef2-b31b-48cd-8577-7eb3ae6aaca6","Type":"ContainerDied","Data":"8a2b033564cdebddca3b9be513005c26512c68f558777c706f6a92aa4fd4be00"} Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.570854 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zbj7g" event={"ID":"09726ef2-b31b-48cd-8577-7eb3ae6aaca6","Type":"ContainerDied","Data":"cb46f452e6e568465fc56dc5b372acf5ee5884a9a1a079e3d5b529866a617ad6"} Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.570958 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zbj7g" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.577005 4808 generic.go:334] "Generic (PLEG): container finished" podID="02097ffc-aa46-438f-b11d-c98301f3e617" containerID="c4917adf048a4ec961ca2f6c21b80a0dddab2f112a5bc63ba2a21834b63ce6cc" exitCode=0 Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.577170 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" event={"ID":"02097ffc-aa46-438f-b11d-c98301f3e617","Type":"ContainerDied","Data":"c4917adf048a4ec961ca2f6c21b80a0dddab2f112a5bc63ba2a21834b63ce6cc"} Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.577295 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" event={"ID":"02097ffc-aa46-438f-b11d-c98301f3e617","Type":"ContainerDied","Data":"fb52e103568780b100651c7ec7e80b7c761d63ba21dc16dd926c6b13d584b060"} Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.577409 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gpxn2" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.587643 4808 generic.go:334] "Generic (PLEG): container finished" podID="21467289-5a33-43ed-ac9b-425774019de0" containerID="b3043c438e926bf5a56d3705fe133d1d20c3d23a69a9b5ef67478bc06c8fef65" exitCode=0 Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.587728 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dz6nh" event={"ID":"21467289-5a33-43ed-ac9b-425774019de0","Type":"ContainerDied","Data":"b3043c438e926bf5a56d3705fe133d1d20c3d23a69a9b5ef67478bc06c8fef65"} Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.587768 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dz6nh" event={"ID":"21467289-5a33-43ed-ac9b-425774019de0","Type":"ContainerDied","Data":"1f347099fc8b55be1ed5172e26616a8e60afbdd2547dd27e07011f3999bab1d9"} Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.587834 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dz6nh" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.591701 4808 scope.go:117] "RemoveContainer" containerID="5ad2f3ab411e52716530e2532e0b72602313552c4427d5baf112290ca04ea400" Oct 02 16:44:53 crc kubenswrapper[4808]: W1002 16:44:53.592121 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod037881d6_84ee_47ba_a0cf_f6ac68f7ef17.slice/crio-f87302e1c40a5eafc16cb02c12a5c86004cad0dde387874f3fd98780e4d33168 WatchSource:0}: Error finding container f87302e1c40a5eafc16cb02c12a5c86004cad0dde387874f3fd98780e4d33168: Status 404 returned error can't find the container with id f87302e1c40a5eafc16cb02c12a5c86004cad0dde387874f3fd98780e4d33168 Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.609468 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhxtz"] Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.610701 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21467289-5a33-43ed-ac9b-425774019de0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.616390 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhxtz"] Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.619574 4808 scope.go:117] "RemoveContainer" containerID="e18ca276d6a428f6592fba1b8d263b238a8e25dfba04e6cde28de8eb28d55590" Oct 02 16:44:53 crc kubenswrapper[4808]: E1002 16:44:53.620330 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e18ca276d6a428f6592fba1b8d263b238a8e25dfba04e6cde28de8eb28d55590\": container with ID starting with e18ca276d6a428f6592fba1b8d263b238a8e25dfba04e6cde28de8eb28d55590 not found: ID does not exist" containerID="e18ca276d6a428f6592fba1b8d263b238a8e25dfba04e6cde28de8eb28d55590" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.620380 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e18ca276d6a428f6592fba1b8d263b238a8e25dfba04e6cde28de8eb28d55590"} err="failed to get container status \"e18ca276d6a428f6592fba1b8d263b238a8e25dfba04e6cde28de8eb28d55590\": rpc error: code = NotFound desc = could not find container \"e18ca276d6a428f6592fba1b8d263b238a8e25dfba04e6cde28de8eb28d55590\": container with ID starting with e18ca276d6a428f6592fba1b8d263b238a8e25dfba04e6cde28de8eb28d55590 not found: ID does not exist" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.620415 4808 scope.go:117] "RemoveContainer" containerID="2d0cef968392141b5021942cd9c8ea255ef06c79aa80c0ba9b3a1139fb2cc605" Oct 02 16:44:53 crc kubenswrapper[4808]: E1002 16:44:53.622041 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d0cef968392141b5021942cd9c8ea255ef06c79aa80c0ba9b3a1139fb2cc605\": container with ID starting with 2d0cef968392141b5021942cd9c8ea255ef06c79aa80c0ba9b3a1139fb2cc605 not found: ID does not exist" containerID="2d0cef968392141b5021942cd9c8ea255ef06c79aa80c0ba9b3a1139fb2cc605" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.622091 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d0cef968392141b5021942cd9c8ea255ef06c79aa80c0ba9b3a1139fb2cc605"} err="failed to get container status \"2d0cef968392141b5021942cd9c8ea255ef06c79aa80c0ba9b3a1139fb2cc605\": rpc error: code = NotFound desc = could not find container \"2d0cef968392141b5021942cd9c8ea255ef06c79aa80c0ba9b3a1139fb2cc605\": container with ID starting with 2d0cef968392141b5021942cd9c8ea255ef06c79aa80c0ba9b3a1139fb2cc605 not found: ID does not exist" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.622126 4808 scope.go:117] "RemoveContainer" containerID="5ad2f3ab411e52716530e2532e0b72602313552c4427d5baf112290ca04ea400" Oct 02 16:44:53 crc kubenswrapper[4808]: E1002 16:44:53.623350 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ad2f3ab411e52716530e2532e0b72602313552c4427d5baf112290ca04ea400\": container with ID starting with 5ad2f3ab411e52716530e2532e0b72602313552c4427d5baf112290ca04ea400 not found: ID does not exist" containerID="5ad2f3ab411e52716530e2532e0b72602313552c4427d5baf112290ca04ea400" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.623396 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ad2f3ab411e52716530e2532e0b72602313552c4427d5baf112290ca04ea400"} err="failed to get container status \"5ad2f3ab411e52716530e2532e0b72602313552c4427d5baf112290ca04ea400\": rpc error: code = NotFound desc = could not find container \"5ad2f3ab411e52716530e2532e0b72602313552c4427d5baf112290ca04ea400\": container with ID starting with 5ad2f3ab411e52716530e2532e0b72602313552c4427d5baf112290ca04ea400 not found: ID does not exist" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.623416 4808 scope.go:117] "RemoveContainer" containerID="b9a880c7820137fd91e6b6a3bd59e14b8a4f4125675e748789d074caaa3a76c3" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.628557 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zbj7g"] Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.640554 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zbj7g"] Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.641751 4808 scope.go:117] "RemoveContainer" containerID="464c5a518d670cf7c0a3bb5d7ede85bf8f86a8fa1b72e109e2a3d000b5f4f028" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.643806 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gpxn2"] Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.646135 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gpxn2"] Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.648438 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nzsf5"] Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.650640 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nzsf5"] Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.669696 4808 scope.go:117] "RemoveContainer" containerID="df0b439bf0b7cac24879d695e54118304faeecc8026be631c1d497c428de9b4d" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.672972 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dz6nh"] Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.678755 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dz6nh"] Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.702979 4808 scope.go:117] "RemoveContainer" containerID="b9a880c7820137fd91e6b6a3bd59e14b8a4f4125675e748789d074caaa3a76c3" Oct 02 16:44:53 crc kubenswrapper[4808]: E1002 16:44:53.703584 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9a880c7820137fd91e6b6a3bd59e14b8a4f4125675e748789d074caaa3a76c3\": container with ID starting with b9a880c7820137fd91e6b6a3bd59e14b8a4f4125675e748789d074caaa3a76c3 not found: ID does not exist" containerID="b9a880c7820137fd91e6b6a3bd59e14b8a4f4125675e748789d074caaa3a76c3" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.703619 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9a880c7820137fd91e6b6a3bd59e14b8a4f4125675e748789d074caaa3a76c3"} err="failed to get container status \"b9a880c7820137fd91e6b6a3bd59e14b8a4f4125675e748789d074caaa3a76c3\": rpc error: code = NotFound desc = could not find container \"b9a880c7820137fd91e6b6a3bd59e14b8a4f4125675e748789d074caaa3a76c3\": container with ID starting with b9a880c7820137fd91e6b6a3bd59e14b8a4f4125675e748789d074caaa3a76c3 not found: ID does not exist" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.703649 4808 scope.go:117] "RemoveContainer" containerID="464c5a518d670cf7c0a3bb5d7ede85bf8f86a8fa1b72e109e2a3d000b5f4f028" Oct 02 16:44:53 crc kubenswrapper[4808]: E1002 16:44:53.703897 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"464c5a518d670cf7c0a3bb5d7ede85bf8f86a8fa1b72e109e2a3d000b5f4f028\": container with ID starting with 464c5a518d670cf7c0a3bb5d7ede85bf8f86a8fa1b72e109e2a3d000b5f4f028 not found: ID does not exist" containerID="464c5a518d670cf7c0a3bb5d7ede85bf8f86a8fa1b72e109e2a3d000b5f4f028" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.703921 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"464c5a518d670cf7c0a3bb5d7ede85bf8f86a8fa1b72e109e2a3d000b5f4f028"} err="failed to get container status \"464c5a518d670cf7c0a3bb5d7ede85bf8f86a8fa1b72e109e2a3d000b5f4f028\": rpc error: code = NotFound desc = could not find container \"464c5a518d670cf7c0a3bb5d7ede85bf8f86a8fa1b72e109e2a3d000b5f4f028\": container with ID starting with 464c5a518d670cf7c0a3bb5d7ede85bf8f86a8fa1b72e109e2a3d000b5f4f028 not found: ID does not exist" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.703936 4808 scope.go:117] "RemoveContainer" containerID="df0b439bf0b7cac24879d695e54118304faeecc8026be631c1d497c428de9b4d" Oct 02 16:44:53 crc kubenswrapper[4808]: E1002 16:44:53.704304 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df0b439bf0b7cac24879d695e54118304faeecc8026be631c1d497c428de9b4d\": container with ID starting with df0b439bf0b7cac24879d695e54118304faeecc8026be631c1d497c428de9b4d not found: ID does not exist" containerID="df0b439bf0b7cac24879d695e54118304faeecc8026be631c1d497c428de9b4d" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.704325 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df0b439bf0b7cac24879d695e54118304faeecc8026be631c1d497c428de9b4d"} err="failed to get container status \"df0b439bf0b7cac24879d695e54118304faeecc8026be631c1d497c428de9b4d\": rpc error: code = NotFound desc = could not find container \"df0b439bf0b7cac24879d695e54118304faeecc8026be631c1d497c428de9b4d\": container with ID starting with df0b439bf0b7cac24879d695e54118304faeecc8026be631c1d497c428de9b4d not found: ID does not exist" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.704338 4808 scope.go:117] "RemoveContainer" containerID="8a2b033564cdebddca3b9be513005c26512c68f558777c706f6a92aa4fd4be00" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.719273 4808 scope.go:117] "RemoveContainer" containerID="1a997847ea15085c0d916c13b2d01c49a434825cf412ae59dda6b36649451caf" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.746619 4808 scope.go:117] "RemoveContainer" containerID="dee5719c4a67f2ea73b66b7b333951b61d6ec163edb4adc147a9cf6980c16e9d" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.760909 4808 scope.go:117] "RemoveContainer" containerID="8a2b033564cdebddca3b9be513005c26512c68f558777c706f6a92aa4fd4be00" Oct 02 16:44:53 crc kubenswrapper[4808]: E1002 16:44:53.762145 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a2b033564cdebddca3b9be513005c26512c68f558777c706f6a92aa4fd4be00\": container with ID starting with 8a2b033564cdebddca3b9be513005c26512c68f558777c706f6a92aa4fd4be00 not found: ID does not exist" containerID="8a2b033564cdebddca3b9be513005c26512c68f558777c706f6a92aa4fd4be00" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.762179 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a2b033564cdebddca3b9be513005c26512c68f558777c706f6a92aa4fd4be00"} err="failed to get container status \"8a2b033564cdebddca3b9be513005c26512c68f558777c706f6a92aa4fd4be00\": rpc error: code = NotFound desc = could not find container \"8a2b033564cdebddca3b9be513005c26512c68f558777c706f6a92aa4fd4be00\": container with ID starting with 8a2b033564cdebddca3b9be513005c26512c68f558777c706f6a92aa4fd4be00 not found: ID does not exist" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.762222 4808 scope.go:117] "RemoveContainer" containerID="1a997847ea15085c0d916c13b2d01c49a434825cf412ae59dda6b36649451caf" Oct 02 16:44:53 crc kubenswrapper[4808]: E1002 16:44:53.762698 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a997847ea15085c0d916c13b2d01c49a434825cf412ae59dda6b36649451caf\": container with ID starting with 1a997847ea15085c0d916c13b2d01c49a434825cf412ae59dda6b36649451caf not found: ID does not exist" containerID="1a997847ea15085c0d916c13b2d01c49a434825cf412ae59dda6b36649451caf" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.762720 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a997847ea15085c0d916c13b2d01c49a434825cf412ae59dda6b36649451caf"} err="failed to get container status \"1a997847ea15085c0d916c13b2d01c49a434825cf412ae59dda6b36649451caf\": rpc error: code = NotFound desc = could not find container \"1a997847ea15085c0d916c13b2d01c49a434825cf412ae59dda6b36649451caf\": container with ID starting with 1a997847ea15085c0d916c13b2d01c49a434825cf412ae59dda6b36649451caf not found: ID does not exist" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.762734 4808 scope.go:117] "RemoveContainer" containerID="dee5719c4a67f2ea73b66b7b333951b61d6ec163edb4adc147a9cf6980c16e9d" Oct 02 16:44:53 crc kubenswrapper[4808]: E1002 16:44:53.763276 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dee5719c4a67f2ea73b66b7b333951b61d6ec163edb4adc147a9cf6980c16e9d\": container with ID starting with dee5719c4a67f2ea73b66b7b333951b61d6ec163edb4adc147a9cf6980c16e9d not found: ID does not exist" containerID="dee5719c4a67f2ea73b66b7b333951b61d6ec163edb4adc147a9cf6980c16e9d" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.763295 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dee5719c4a67f2ea73b66b7b333951b61d6ec163edb4adc147a9cf6980c16e9d"} err="failed to get container status \"dee5719c4a67f2ea73b66b7b333951b61d6ec163edb4adc147a9cf6980c16e9d\": rpc error: code = NotFound desc = could not find container \"dee5719c4a67f2ea73b66b7b333951b61d6ec163edb4adc147a9cf6980c16e9d\": container with ID starting with dee5719c4a67f2ea73b66b7b333951b61d6ec163edb4adc147a9cf6980c16e9d not found: ID does not exist" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.763311 4808 scope.go:117] "RemoveContainer" containerID="c4917adf048a4ec961ca2f6c21b80a0dddab2f112a5bc63ba2a21834b63ce6cc" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.780735 4808 scope.go:117] "RemoveContainer" containerID="c4917adf048a4ec961ca2f6c21b80a0dddab2f112a5bc63ba2a21834b63ce6cc" Oct 02 16:44:53 crc kubenswrapper[4808]: E1002 16:44:53.781439 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4917adf048a4ec961ca2f6c21b80a0dddab2f112a5bc63ba2a21834b63ce6cc\": container with ID starting with c4917adf048a4ec961ca2f6c21b80a0dddab2f112a5bc63ba2a21834b63ce6cc not found: ID does not exist" containerID="c4917adf048a4ec961ca2f6c21b80a0dddab2f112a5bc63ba2a21834b63ce6cc" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.781489 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4917adf048a4ec961ca2f6c21b80a0dddab2f112a5bc63ba2a21834b63ce6cc"} err="failed to get container status \"c4917adf048a4ec961ca2f6c21b80a0dddab2f112a5bc63ba2a21834b63ce6cc\": rpc error: code = NotFound desc = could not find container \"c4917adf048a4ec961ca2f6c21b80a0dddab2f112a5bc63ba2a21834b63ce6cc\": container with ID starting with c4917adf048a4ec961ca2f6c21b80a0dddab2f112a5bc63ba2a21834b63ce6cc not found: ID does not exist" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.781510 4808 scope.go:117] "RemoveContainer" containerID="b3043c438e926bf5a56d3705fe133d1d20c3d23a69a9b5ef67478bc06c8fef65" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.803690 4808 scope.go:117] "RemoveContainer" containerID="6398818fd73d7f37e43e21bde66a7e1f69fc9b85f4cec1ab92b1526af73d4fb8" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.823271 4808 scope.go:117] "RemoveContainer" containerID="5e26b51042540c6addba14edd250706f5f348432b9ef6566b02aa7ac81c2fd97" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.842017 4808 scope.go:117] "RemoveContainer" containerID="b3043c438e926bf5a56d3705fe133d1d20c3d23a69a9b5ef67478bc06c8fef65" Oct 02 16:44:53 crc kubenswrapper[4808]: E1002 16:44:53.842895 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3043c438e926bf5a56d3705fe133d1d20c3d23a69a9b5ef67478bc06c8fef65\": container with ID starting with b3043c438e926bf5a56d3705fe133d1d20c3d23a69a9b5ef67478bc06c8fef65 not found: ID does not exist" containerID="b3043c438e926bf5a56d3705fe133d1d20c3d23a69a9b5ef67478bc06c8fef65" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.842971 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3043c438e926bf5a56d3705fe133d1d20c3d23a69a9b5ef67478bc06c8fef65"} err="failed to get container status \"b3043c438e926bf5a56d3705fe133d1d20c3d23a69a9b5ef67478bc06c8fef65\": rpc error: code = NotFound desc = could not find container \"b3043c438e926bf5a56d3705fe133d1d20c3d23a69a9b5ef67478bc06c8fef65\": container with ID starting with b3043c438e926bf5a56d3705fe133d1d20c3d23a69a9b5ef67478bc06c8fef65 not found: ID does not exist" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.843022 4808 scope.go:117] "RemoveContainer" containerID="6398818fd73d7f37e43e21bde66a7e1f69fc9b85f4cec1ab92b1526af73d4fb8" Oct 02 16:44:53 crc kubenswrapper[4808]: E1002 16:44:53.843589 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6398818fd73d7f37e43e21bde66a7e1f69fc9b85f4cec1ab92b1526af73d4fb8\": container with ID starting with 6398818fd73d7f37e43e21bde66a7e1f69fc9b85f4cec1ab92b1526af73d4fb8 not found: ID does not exist" containerID="6398818fd73d7f37e43e21bde66a7e1f69fc9b85f4cec1ab92b1526af73d4fb8" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.843633 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6398818fd73d7f37e43e21bde66a7e1f69fc9b85f4cec1ab92b1526af73d4fb8"} err="failed to get container status \"6398818fd73d7f37e43e21bde66a7e1f69fc9b85f4cec1ab92b1526af73d4fb8\": rpc error: code = NotFound desc = could not find container \"6398818fd73d7f37e43e21bde66a7e1f69fc9b85f4cec1ab92b1526af73d4fb8\": container with ID starting with 6398818fd73d7f37e43e21bde66a7e1f69fc9b85f4cec1ab92b1526af73d4fb8 not found: ID does not exist" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.843666 4808 scope.go:117] "RemoveContainer" containerID="5e26b51042540c6addba14edd250706f5f348432b9ef6566b02aa7ac81c2fd97" Oct 02 16:44:53 crc kubenswrapper[4808]: E1002 16:44:53.844279 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e26b51042540c6addba14edd250706f5f348432b9ef6566b02aa7ac81c2fd97\": container with ID starting with 5e26b51042540c6addba14edd250706f5f348432b9ef6566b02aa7ac81c2fd97 not found: ID does not exist" containerID="5e26b51042540c6addba14edd250706f5f348432b9ef6566b02aa7ac81c2fd97" Oct 02 16:44:53 crc kubenswrapper[4808]: I1002 16:44:53.844355 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e26b51042540c6addba14edd250706f5f348432b9ef6566b02aa7ac81c2fd97"} err="failed to get container status \"5e26b51042540c6addba14edd250706f5f348432b9ef6566b02aa7ac81c2fd97\": rpc error: code = NotFound desc = could not find container \"5e26b51042540c6addba14edd250706f5f348432b9ef6566b02aa7ac81c2fd97\": container with ID starting with 5e26b51042540c6addba14edd250706f5f348432b9ef6566b02aa7ac81c2fd97 not found: ID does not exist" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.221179 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8974h"] Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.596719 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bhdpb" event={"ID":"037881d6-84ee-47ba-a0cf-f6ac68f7ef17","Type":"ContainerStarted","Data":"f08197ff1a62f6646b1bde37686ae07ec9f7c47c2aa7ae61cc17b1adef0bc027"} Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.597086 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bhdpb" event={"ID":"037881d6-84ee-47ba-a0cf-f6ac68f7ef17","Type":"ContainerStarted","Data":"f87302e1c40a5eafc16cb02c12a5c86004cad0dde387874f3fd98780e4d33168"} Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.598593 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-bhdpb" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.602005 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-bhdpb" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.617431 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-bhdpb" podStartSLOduration=2.6174070929999997 podStartE2EDuration="2.617407093s" podCreationTimestamp="2025-10-02 16:44:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:44:54.61489342 +0000 UTC m=+261.940422420" watchObservedRunningTime="2025-10-02 16:44:54.617407093 +0000 UTC m=+261.942936093" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.793488 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nbjkh"] Oct 02 16:44:54 crc kubenswrapper[4808]: E1002 16:44:54.793818 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c0f4487-6602-4c9d-b414-229588b2e797" containerName="extract-utilities" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.793840 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c0f4487-6602-4c9d-b414-229588b2e797" containerName="extract-utilities" Oct 02 16:44:54 crc kubenswrapper[4808]: E1002 16:44:54.793862 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c0f4487-6602-4c9d-b414-229588b2e797" containerName="extract-content" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.793878 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c0f4487-6602-4c9d-b414-229588b2e797" containerName="extract-content" Oct 02 16:44:54 crc kubenswrapper[4808]: E1002 16:44:54.793892 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21467289-5a33-43ed-ac9b-425774019de0" containerName="extract-utilities" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.793908 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="21467289-5a33-43ed-ac9b-425774019de0" containerName="extract-utilities" Oct 02 16:44:54 crc kubenswrapper[4808]: E1002 16:44:54.793929 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09726ef2-b31b-48cd-8577-7eb3ae6aaca6" containerName="extract-content" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.793941 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="09726ef2-b31b-48cd-8577-7eb3ae6aaca6" containerName="extract-content" Oct 02 16:44:54 crc kubenswrapper[4808]: E1002 16:44:54.793956 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa0909cd-83b3-4127-abc6-2540f61e0b2b" containerName="extract-content" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.793968 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa0909cd-83b3-4127-abc6-2540f61e0b2b" containerName="extract-content" Oct 02 16:44:54 crc kubenswrapper[4808]: E1002 16:44:54.793988 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02097ffc-aa46-438f-b11d-c98301f3e617" containerName="marketplace-operator" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.794000 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="02097ffc-aa46-438f-b11d-c98301f3e617" containerName="marketplace-operator" Oct 02 16:44:54 crc kubenswrapper[4808]: E1002 16:44:54.794021 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21467289-5a33-43ed-ac9b-425774019de0" containerName="registry-server" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.794033 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="21467289-5a33-43ed-ac9b-425774019de0" containerName="registry-server" Oct 02 16:44:54 crc kubenswrapper[4808]: E1002 16:44:54.794048 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09726ef2-b31b-48cd-8577-7eb3ae6aaca6" containerName="registry-server" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.794061 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="09726ef2-b31b-48cd-8577-7eb3ae6aaca6" containerName="registry-server" Oct 02 16:44:54 crc kubenswrapper[4808]: E1002 16:44:54.794077 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21467289-5a33-43ed-ac9b-425774019de0" containerName="extract-content" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.794089 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="21467289-5a33-43ed-ac9b-425774019de0" containerName="extract-content" Oct 02 16:44:54 crc kubenswrapper[4808]: E1002 16:44:54.794111 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c0f4487-6602-4c9d-b414-229588b2e797" containerName="registry-server" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.794123 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c0f4487-6602-4c9d-b414-229588b2e797" containerName="registry-server" Oct 02 16:44:54 crc kubenswrapper[4808]: E1002 16:44:54.794138 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09726ef2-b31b-48cd-8577-7eb3ae6aaca6" containerName="extract-utilities" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.794149 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="09726ef2-b31b-48cd-8577-7eb3ae6aaca6" containerName="extract-utilities" Oct 02 16:44:54 crc kubenswrapper[4808]: E1002 16:44:54.794163 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa0909cd-83b3-4127-abc6-2540f61e0b2b" containerName="registry-server" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.794175 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa0909cd-83b3-4127-abc6-2540f61e0b2b" containerName="registry-server" Oct 02 16:44:54 crc kubenswrapper[4808]: E1002 16:44:54.794196 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa0909cd-83b3-4127-abc6-2540f61e0b2b" containerName="extract-utilities" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.794208 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa0909cd-83b3-4127-abc6-2540f61e0b2b" containerName="extract-utilities" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.794382 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="21467289-5a33-43ed-ac9b-425774019de0" containerName="registry-server" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.794411 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa0909cd-83b3-4127-abc6-2540f61e0b2b" containerName="registry-server" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.794425 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="09726ef2-b31b-48cd-8577-7eb3ae6aaca6" containerName="registry-server" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.794443 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c0f4487-6602-4c9d-b414-229588b2e797" containerName="registry-server" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.794460 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="02097ffc-aa46-438f-b11d-c98301f3e617" containerName="marketplace-operator" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.795688 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nbjkh" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.801174 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.803221 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nbjkh"] Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.830049 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6893a3db-ff37-480f-9883-af8db609de88-catalog-content\") pod \"certified-operators-nbjkh\" (UID: \"6893a3db-ff37-480f-9883-af8db609de88\") " pod="openshift-marketplace/certified-operators-nbjkh" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.830111 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6893a3db-ff37-480f-9883-af8db609de88-utilities\") pod \"certified-operators-nbjkh\" (UID: \"6893a3db-ff37-480f-9883-af8db609de88\") " pod="openshift-marketplace/certified-operators-nbjkh" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.830151 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzqxg\" (UniqueName: \"kubernetes.io/projected/6893a3db-ff37-480f-9883-af8db609de88-kube-api-access-fzqxg\") pod \"certified-operators-nbjkh\" (UID: \"6893a3db-ff37-480f-9883-af8db609de88\") " pod="openshift-marketplace/certified-operators-nbjkh" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.931881 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6893a3db-ff37-480f-9883-af8db609de88-utilities\") pod \"certified-operators-nbjkh\" (UID: \"6893a3db-ff37-480f-9883-af8db609de88\") " pod="openshift-marketplace/certified-operators-nbjkh" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.932827 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6893a3db-ff37-480f-9883-af8db609de88-utilities\") pod \"certified-operators-nbjkh\" (UID: \"6893a3db-ff37-480f-9883-af8db609de88\") " pod="openshift-marketplace/certified-operators-nbjkh" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.933118 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzqxg\" (UniqueName: \"kubernetes.io/projected/6893a3db-ff37-480f-9883-af8db609de88-kube-api-access-fzqxg\") pod \"certified-operators-nbjkh\" (UID: \"6893a3db-ff37-480f-9883-af8db609de88\") " pod="openshift-marketplace/certified-operators-nbjkh" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.933635 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6893a3db-ff37-480f-9883-af8db609de88-catalog-content\") pod \"certified-operators-nbjkh\" (UID: \"6893a3db-ff37-480f-9883-af8db609de88\") " pod="openshift-marketplace/certified-operators-nbjkh" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.933949 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6893a3db-ff37-480f-9883-af8db609de88-catalog-content\") pod \"certified-operators-nbjkh\" (UID: \"6893a3db-ff37-480f-9883-af8db609de88\") " pod="openshift-marketplace/certified-operators-nbjkh" Oct 02 16:44:54 crc kubenswrapper[4808]: I1002 16:44:54.957090 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzqxg\" (UniqueName: \"kubernetes.io/projected/6893a3db-ff37-480f-9883-af8db609de88-kube-api-access-fzqxg\") pod \"certified-operators-nbjkh\" (UID: \"6893a3db-ff37-480f-9883-af8db609de88\") " pod="openshift-marketplace/certified-operators-nbjkh" Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.120671 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nbjkh" Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.380331 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nbjkh"] Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.390559 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k8j2q"] Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.410835 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k8j2q"] Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.413282 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k8j2q" Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.418247 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.418840 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02097ffc-aa46-438f-b11d-c98301f3e617" path="/var/lib/kubelet/pods/02097ffc-aa46-438f-b11d-c98301f3e617/volumes" Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.419845 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09726ef2-b31b-48cd-8577-7eb3ae6aaca6" path="/var/lib/kubelet/pods/09726ef2-b31b-48cd-8577-7eb3ae6aaca6/volumes" Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.420601 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21467289-5a33-43ed-ac9b-425774019de0" path="/var/lib/kubelet/pods/21467289-5a33-43ed-ac9b-425774019de0/volumes" Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.421798 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c0f4487-6602-4c9d-b414-229588b2e797" path="/var/lib/kubelet/pods/6c0f4487-6602-4c9d-b414-229588b2e797/volumes" Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.431733 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa0909cd-83b3-4127-abc6-2540f61e0b2b" path="/var/lib/kubelet/pods/fa0909cd-83b3-4127-abc6-2540f61e0b2b/volumes" Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.442654 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8tr8\" (UniqueName: \"kubernetes.io/projected/3f88e6e2-0f03-4f48-8040-b598130fa460-kube-api-access-b8tr8\") pod \"redhat-marketplace-k8j2q\" (UID: \"3f88e6e2-0f03-4f48-8040-b598130fa460\") " pod="openshift-marketplace/redhat-marketplace-k8j2q" Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.442750 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f88e6e2-0f03-4f48-8040-b598130fa460-catalog-content\") pod \"redhat-marketplace-k8j2q\" (UID: \"3f88e6e2-0f03-4f48-8040-b598130fa460\") " pod="openshift-marketplace/redhat-marketplace-k8j2q" Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.442820 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f88e6e2-0f03-4f48-8040-b598130fa460-utilities\") pod \"redhat-marketplace-k8j2q\" (UID: \"3f88e6e2-0f03-4f48-8040-b598130fa460\") " pod="openshift-marketplace/redhat-marketplace-k8j2q" Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.544211 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8tr8\" (UniqueName: \"kubernetes.io/projected/3f88e6e2-0f03-4f48-8040-b598130fa460-kube-api-access-b8tr8\") pod \"redhat-marketplace-k8j2q\" (UID: \"3f88e6e2-0f03-4f48-8040-b598130fa460\") " pod="openshift-marketplace/redhat-marketplace-k8j2q" Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.545614 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f88e6e2-0f03-4f48-8040-b598130fa460-catalog-content\") pod \"redhat-marketplace-k8j2q\" (UID: \"3f88e6e2-0f03-4f48-8040-b598130fa460\") " pod="openshift-marketplace/redhat-marketplace-k8j2q" Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.546088 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f88e6e2-0f03-4f48-8040-b598130fa460-catalog-content\") pod \"redhat-marketplace-k8j2q\" (UID: \"3f88e6e2-0f03-4f48-8040-b598130fa460\") " pod="openshift-marketplace/redhat-marketplace-k8j2q" Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.546334 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f88e6e2-0f03-4f48-8040-b598130fa460-utilities\") pod \"redhat-marketplace-k8j2q\" (UID: \"3f88e6e2-0f03-4f48-8040-b598130fa460\") " pod="openshift-marketplace/redhat-marketplace-k8j2q" Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.546640 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f88e6e2-0f03-4f48-8040-b598130fa460-utilities\") pod \"redhat-marketplace-k8j2q\" (UID: \"3f88e6e2-0f03-4f48-8040-b598130fa460\") " pod="openshift-marketplace/redhat-marketplace-k8j2q" Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.567048 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8tr8\" (UniqueName: \"kubernetes.io/projected/3f88e6e2-0f03-4f48-8040-b598130fa460-kube-api-access-b8tr8\") pod \"redhat-marketplace-k8j2q\" (UID: \"3f88e6e2-0f03-4f48-8040-b598130fa460\") " pod="openshift-marketplace/redhat-marketplace-k8j2q" Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.610624 4808 generic.go:334] "Generic (PLEG): container finished" podID="6893a3db-ff37-480f-9883-af8db609de88" containerID="57b1a308b3e3de6490b7c104fc27da11a612327d2d50f5e7b3633cbadf3ff9c2" exitCode=0 Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.611348 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nbjkh" event={"ID":"6893a3db-ff37-480f-9883-af8db609de88","Type":"ContainerDied","Data":"57b1a308b3e3de6490b7c104fc27da11a612327d2d50f5e7b3633cbadf3ff9c2"} Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.611375 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nbjkh" event={"ID":"6893a3db-ff37-480f-9883-af8db609de88","Type":"ContainerStarted","Data":"451dab5d68c99dd99f18913296a4f09916ee2d08f24533a9d0dd66e7574d8de9"} Oct 02 16:44:55 crc kubenswrapper[4808]: I1002 16:44:55.787282 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k8j2q" Oct 02 16:44:56 crc kubenswrapper[4808]: I1002 16:44:56.025301 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k8j2q"] Oct 02 16:44:56 crc kubenswrapper[4808]: I1002 16:44:56.618928 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nbjkh" event={"ID":"6893a3db-ff37-480f-9883-af8db609de88","Type":"ContainerStarted","Data":"5732cc48d7fa41e2d264fcb515ea76db9c7df3cad046512a64b0de5e81874379"} Oct 02 16:44:56 crc kubenswrapper[4808]: I1002 16:44:56.620851 4808 generic.go:334] "Generic (PLEG): container finished" podID="3f88e6e2-0f03-4f48-8040-b598130fa460" containerID="6b801dae8603cf336de12cce4f5df90ac29a1e286f4e98576bd240101e116045" exitCode=0 Oct 02 16:44:56 crc kubenswrapper[4808]: I1002 16:44:56.620909 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k8j2q" event={"ID":"3f88e6e2-0f03-4f48-8040-b598130fa460","Type":"ContainerDied","Data":"6b801dae8603cf336de12cce4f5df90ac29a1e286f4e98576bd240101e116045"} Oct 02 16:44:56 crc kubenswrapper[4808]: I1002 16:44:56.620997 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k8j2q" event={"ID":"3f88e6e2-0f03-4f48-8040-b598130fa460","Type":"ContainerStarted","Data":"1e84a77c15bc79cf16d2876e7e788d8b73edc3013f9265c2ed133177aa6abbc0"} Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.203457 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n6gnm"] Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.205024 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n6gnm" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.208292 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.216726 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n6gnm"] Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.273870 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3057619c-ad06-40e4-905b-981c2f7cf6d4-utilities\") pod \"redhat-operators-n6gnm\" (UID: \"3057619c-ad06-40e4-905b-981c2f7cf6d4\") " pod="openshift-marketplace/redhat-operators-n6gnm" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.274529 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tzb5\" (UniqueName: \"kubernetes.io/projected/3057619c-ad06-40e4-905b-981c2f7cf6d4-kube-api-access-5tzb5\") pod \"redhat-operators-n6gnm\" (UID: \"3057619c-ad06-40e4-905b-981c2f7cf6d4\") " pod="openshift-marketplace/redhat-operators-n6gnm" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.274816 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3057619c-ad06-40e4-905b-981c2f7cf6d4-catalog-content\") pod \"redhat-operators-n6gnm\" (UID: \"3057619c-ad06-40e4-905b-981c2f7cf6d4\") " pod="openshift-marketplace/redhat-operators-n6gnm" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.376381 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3057619c-ad06-40e4-905b-981c2f7cf6d4-utilities\") pod \"redhat-operators-n6gnm\" (UID: \"3057619c-ad06-40e4-905b-981c2f7cf6d4\") " pod="openshift-marketplace/redhat-operators-n6gnm" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.376444 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tzb5\" (UniqueName: \"kubernetes.io/projected/3057619c-ad06-40e4-905b-981c2f7cf6d4-kube-api-access-5tzb5\") pod \"redhat-operators-n6gnm\" (UID: \"3057619c-ad06-40e4-905b-981c2f7cf6d4\") " pod="openshift-marketplace/redhat-operators-n6gnm" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.376502 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3057619c-ad06-40e4-905b-981c2f7cf6d4-catalog-content\") pod \"redhat-operators-n6gnm\" (UID: \"3057619c-ad06-40e4-905b-981c2f7cf6d4\") " pod="openshift-marketplace/redhat-operators-n6gnm" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.377016 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3057619c-ad06-40e4-905b-981c2f7cf6d4-catalog-content\") pod \"redhat-operators-n6gnm\" (UID: \"3057619c-ad06-40e4-905b-981c2f7cf6d4\") " pod="openshift-marketplace/redhat-operators-n6gnm" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.377721 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3057619c-ad06-40e4-905b-981c2f7cf6d4-utilities\") pod \"redhat-operators-n6gnm\" (UID: \"3057619c-ad06-40e4-905b-981c2f7cf6d4\") " pod="openshift-marketplace/redhat-operators-n6gnm" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.405870 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tzb5\" (UniqueName: \"kubernetes.io/projected/3057619c-ad06-40e4-905b-981c2f7cf6d4-kube-api-access-5tzb5\") pod \"redhat-operators-n6gnm\" (UID: \"3057619c-ad06-40e4-905b-981c2f7cf6d4\") " pod="openshift-marketplace/redhat-operators-n6gnm" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.529304 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n6gnm" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.631373 4808 generic.go:334] "Generic (PLEG): container finished" podID="6893a3db-ff37-480f-9883-af8db609de88" containerID="5732cc48d7fa41e2d264fcb515ea76db9c7df3cad046512a64b0de5e81874379" exitCode=0 Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.631431 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nbjkh" event={"ID":"6893a3db-ff37-480f-9883-af8db609de88","Type":"ContainerDied","Data":"5732cc48d7fa41e2d264fcb515ea76db9c7df3cad046512a64b0de5e81874379"} Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.635545 4808 generic.go:334] "Generic (PLEG): container finished" podID="3f88e6e2-0f03-4f48-8040-b598130fa460" containerID="a811c5cafa292417d0a469f7a77bc5a99c9d22185e586c3e5ab023627772f85a" exitCode=0 Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.635600 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k8j2q" event={"ID":"3f88e6e2-0f03-4f48-8040-b598130fa460","Type":"ContainerDied","Data":"a811c5cafa292417d0a469f7a77bc5a99c9d22185e586c3e5ab023627772f85a"} Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.789752 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4q65f"] Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.791098 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4q65f" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.793580 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.799379 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4q65f"] Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.885332 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7ae49e8-10fa-4144-aa62-30d63a77ab0e-utilities\") pod \"community-operators-4q65f\" (UID: \"b7ae49e8-10fa-4144-aa62-30d63a77ab0e\") " pod="openshift-marketplace/community-operators-4q65f" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.885395 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7ae49e8-10fa-4144-aa62-30d63a77ab0e-catalog-content\") pod \"community-operators-4q65f\" (UID: \"b7ae49e8-10fa-4144-aa62-30d63a77ab0e\") " pod="openshift-marketplace/community-operators-4q65f" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.885437 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6rdr\" (UniqueName: \"kubernetes.io/projected/b7ae49e8-10fa-4144-aa62-30d63a77ab0e-kube-api-access-m6rdr\") pod \"community-operators-4q65f\" (UID: \"b7ae49e8-10fa-4144-aa62-30d63a77ab0e\") " pod="openshift-marketplace/community-operators-4q65f" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.957523 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n6gnm"] Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.987600 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7ae49e8-10fa-4144-aa62-30d63a77ab0e-utilities\") pod \"community-operators-4q65f\" (UID: \"b7ae49e8-10fa-4144-aa62-30d63a77ab0e\") " pod="openshift-marketplace/community-operators-4q65f" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.987890 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7ae49e8-10fa-4144-aa62-30d63a77ab0e-catalog-content\") pod \"community-operators-4q65f\" (UID: \"b7ae49e8-10fa-4144-aa62-30d63a77ab0e\") " pod="openshift-marketplace/community-operators-4q65f" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.988030 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6rdr\" (UniqueName: \"kubernetes.io/projected/b7ae49e8-10fa-4144-aa62-30d63a77ab0e-kube-api-access-m6rdr\") pod \"community-operators-4q65f\" (UID: \"b7ae49e8-10fa-4144-aa62-30d63a77ab0e\") " pod="openshift-marketplace/community-operators-4q65f" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.988322 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7ae49e8-10fa-4144-aa62-30d63a77ab0e-utilities\") pod \"community-operators-4q65f\" (UID: \"b7ae49e8-10fa-4144-aa62-30d63a77ab0e\") " pod="openshift-marketplace/community-operators-4q65f" Oct 02 16:44:57 crc kubenswrapper[4808]: I1002 16:44:57.989446 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7ae49e8-10fa-4144-aa62-30d63a77ab0e-catalog-content\") pod \"community-operators-4q65f\" (UID: \"b7ae49e8-10fa-4144-aa62-30d63a77ab0e\") " pod="openshift-marketplace/community-operators-4q65f" Oct 02 16:44:58 crc kubenswrapper[4808]: I1002 16:44:58.007209 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6rdr\" (UniqueName: \"kubernetes.io/projected/b7ae49e8-10fa-4144-aa62-30d63a77ab0e-kube-api-access-m6rdr\") pod \"community-operators-4q65f\" (UID: \"b7ae49e8-10fa-4144-aa62-30d63a77ab0e\") " pod="openshift-marketplace/community-operators-4q65f" Oct 02 16:44:58 crc kubenswrapper[4808]: I1002 16:44:58.115310 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4q65f" Oct 02 16:44:58 crc kubenswrapper[4808]: I1002 16:44:58.645560 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nbjkh" event={"ID":"6893a3db-ff37-480f-9883-af8db609de88","Type":"ContainerStarted","Data":"e9d0daeba70009d61c4de0a7d458cb23924211076fffd7b66abacde42c07d1bb"} Oct 02 16:44:58 crc kubenswrapper[4808]: I1002 16:44:58.649355 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k8j2q" event={"ID":"3f88e6e2-0f03-4f48-8040-b598130fa460","Type":"ContainerStarted","Data":"cd3f9017e276ee09e046c67194df294edbf89bde2d7c0e88d17f50c73bcaa05d"} Oct 02 16:44:58 crc kubenswrapper[4808]: I1002 16:44:58.651996 4808 generic.go:334] "Generic (PLEG): container finished" podID="3057619c-ad06-40e4-905b-981c2f7cf6d4" containerID="e38f9fb421f53f682b70faae304238cc7a32ed5a3ca0f21a43b6a09ae1a7dea1" exitCode=0 Oct 02 16:44:58 crc kubenswrapper[4808]: I1002 16:44:58.652031 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6gnm" event={"ID":"3057619c-ad06-40e4-905b-981c2f7cf6d4","Type":"ContainerDied","Data":"e38f9fb421f53f682b70faae304238cc7a32ed5a3ca0f21a43b6a09ae1a7dea1"} Oct 02 16:44:58 crc kubenswrapper[4808]: I1002 16:44:58.652052 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6gnm" event={"ID":"3057619c-ad06-40e4-905b-981c2f7cf6d4","Type":"ContainerStarted","Data":"5df466f626e12682fa05021d71a90b1cb0ee26235475ea6aeb95286a96efee6b"} Oct 02 16:44:58 crc kubenswrapper[4808]: I1002 16:44:58.667029 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nbjkh" podStartSLOduration=2.011619995 podStartE2EDuration="4.667004867s" podCreationTimestamp="2025-10-02 16:44:54 +0000 UTC" firstStartedPulling="2025-10-02 16:44:55.612064622 +0000 UTC m=+262.937593622" lastFinishedPulling="2025-10-02 16:44:58.267449494 +0000 UTC m=+265.592978494" observedRunningTime="2025-10-02 16:44:58.664956653 +0000 UTC m=+265.990485673" watchObservedRunningTime="2025-10-02 16:44:58.667004867 +0000 UTC m=+265.992533867" Oct 02 16:44:58 crc kubenswrapper[4808]: I1002 16:44:58.710020 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k8j2q" podStartSLOduration=2.103634511 podStartE2EDuration="3.709995874s" podCreationTimestamp="2025-10-02 16:44:55 +0000 UTC" firstStartedPulling="2025-10-02 16:44:56.622286605 +0000 UTC m=+263.947815605" lastFinishedPulling="2025-10-02 16:44:58.228647968 +0000 UTC m=+265.554176968" observedRunningTime="2025-10-02 16:44:58.706126856 +0000 UTC m=+266.031655866" watchObservedRunningTime="2025-10-02 16:44:58.709995874 +0000 UTC m=+266.035524874" Oct 02 16:44:58 crc kubenswrapper[4808]: I1002 16:44:58.735671 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4q65f"] Oct 02 16:44:58 crc kubenswrapper[4808]: W1002 16:44:58.748506 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7ae49e8_10fa_4144_aa62_30d63a77ab0e.slice/crio-16f59a44abe80faeced28d7beb57f5b46e1ae1d3753ca86c81d61d44ac020a6a WatchSource:0}: Error finding container 16f59a44abe80faeced28d7beb57f5b46e1ae1d3753ca86c81d61d44ac020a6a: Status 404 returned error can't find the container with id 16f59a44abe80faeced28d7beb57f5b46e1ae1d3753ca86c81d61d44ac020a6a Oct 02 16:44:59 crc kubenswrapper[4808]: I1002 16:44:59.660434 4808 generic.go:334] "Generic (PLEG): container finished" podID="b7ae49e8-10fa-4144-aa62-30d63a77ab0e" containerID="9aa3c511a0c211d8b78e932ff169040bec080a80ab3c8395d146d502a2664deb" exitCode=0 Oct 02 16:44:59 crc kubenswrapper[4808]: I1002 16:44:59.660970 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4q65f" event={"ID":"b7ae49e8-10fa-4144-aa62-30d63a77ab0e","Type":"ContainerDied","Data":"9aa3c511a0c211d8b78e932ff169040bec080a80ab3c8395d146d502a2664deb"} Oct 02 16:44:59 crc kubenswrapper[4808]: I1002 16:44:59.661055 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4q65f" event={"ID":"b7ae49e8-10fa-4144-aa62-30d63a77ab0e","Type":"ContainerStarted","Data":"16f59a44abe80faeced28d7beb57f5b46e1ae1d3753ca86c81d61d44ac020a6a"} Oct 02 16:45:00 crc kubenswrapper[4808]: I1002 16:45:00.131775 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd"] Oct 02 16:45:00 crc kubenswrapper[4808]: I1002 16:45:00.132880 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd" Oct 02 16:45:00 crc kubenswrapper[4808]: I1002 16:45:00.135143 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 16:45:00 crc kubenswrapper[4808]: I1002 16:45:00.135489 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 16:45:00 crc kubenswrapper[4808]: I1002 16:45:00.144782 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd"] Oct 02 16:45:00 crc kubenswrapper[4808]: I1002 16:45:00.220183 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/260610f2-c80c-43f5-af6b-b4a4352a1564-secret-volume\") pod \"collect-profiles-29323725-nz4kd\" (UID: \"260610f2-c80c-43f5-af6b-b4a4352a1564\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd" Oct 02 16:45:00 crc kubenswrapper[4808]: I1002 16:45:00.220324 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks2hp\" (UniqueName: \"kubernetes.io/projected/260610f2-c80c-43f5-af6b-b4a4352a1564-kube-api-access-ks2hp\") pod \"collect-profiles-29323725-nz4kd\" (UID: \"260610f2-c80c-43f5-af6b-b4a4352a1564\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd" Oct 02 16:45:00 crc kubenswrapper[4808]: I1002 16:45:00.220359 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/260610f2-c80c-43f5-af6b-b4a4352a1564-config-volume\") pod \"collect-profiles-29323725-nz4kd\" (UID: \"260610f2-c80c-43f5-af6b-b4a4352a1564\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd" Oct 02 16:45:00 crc kubenswrapper[4808]: I1002 16:45:00.321952 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/260610f2-c80c-43f5-af6b-b4a4352a1564-secret-volume\") pod \"collect-profiles-29323725-nz4kd\" (UID: \"260610f2-c80c-43f5-af6b-b4a4352a1564\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd" Oct 02 16:45:00 crc kubenswrapper[4808]: I1002 16:45:00.322053 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks2hp\" (UniqueName: \"kubernetes.io/projected/260610f2-c80c-43f5-af6b-b4a4352a1564-kube-api-access-ks2hp\") pod \"collect-profiles-29323725-nz4kd\" (UID: \"260610f2-c80c-43f5-af6b-b4a4352a1564\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd" Oct 02 16:45:00 crc kubenswrapper[4808]: I1002 16:45:00.322088 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/260610f2-c80c-43f5-af6b-b4a4352a1564-config-volume\") pod \"collect-profiles-29323725-nz4kd\" (UID: \"260610f2-c80c-43f5-af6b-b4a4352a1564\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd" Oct 02 16:45:00 crc kubenswrapper[4808]: I1002 16:45:00.323133 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/260610f2-c80c-43f5-af6b-b4a4352a1564-config-volume\") pod \"collect-profiles-29323725-nz4kd\" (UID: \"260610f2-c80c-43f5-af6b-b4a4352a1564\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd" Oct 02 16:45:00 crc kubenswrapper[4808]: I1002 16:45:00.330558 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/260610f2-c80c-43f5-af6b-b4a4352a1564-secret-volume\") pod \"collect-profiles-29323725-nz4kd\" (UID: \"260610f2-c80c-43f5-af6b-b4a4352a1564\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd" Oct 02 16:45:00 crc kubenswrapper[4808]: I1002 16:45:00.342725 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks2hp\" (UniqueName: \"kubernetes.io/projected/260610f2-c80c-43f5-af6b-b4a4352a1564-kube-api-access-ks2hp\") pod \"collect-profiles-29323725-nz4kd\" (UID: \"260610f2-c80c-43f5-af6b-b4a4352a1564\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd" Oct 02 16:45:00 crc kubenswrapper[4808]: I1002 16:45:00.455656 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd" Oct 02 16:45:00 crc kubenswrapper[4808]: I1002 16:45:00.673869 4808 generic.go:334] "Generic (PLEG): container finished" podID="3057619c-ad06-40e4-905b-981c2f7cf6d4" containerID="13af22bc038059de9af11a4eaf6250b8b61a17419acfdd22c1fd6f5715c036c9" exitCode=0 Oct 02 16:45:00 crc kubenswrapper[4808]: I1002 16:45:00.673924 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6gnm" event={"ID":"3057619c-ad06-40e4-905b-981c2f7cf6d4","Type":"ContainerDied","Data":"13af22bc038059de9af11a4eaf6250b8b61a17419acfdd22c1fd6f5715c036c9"} Oct 02 16:45:00 crc kubenswrapper[4808]: I1002 16:45:00.894951 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd"] Oct 02 16:45:01 crc kubenswrapper[4808]: I1002 16:45:01.682531 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd" event={"ID":"260610f2-c80c-43f5-af6b-b4a4352a1564","Type":"ContainerStarted","Data":"06b3ae2aabb32dc1e7a874e250040a26be3c49061a174e9333c2ee6f30de37d6"} Oct 02 16:45:02 crc kubenswrapper[4808]: I1002 16:45:02.692036 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6gnm" event={"ID":"3057619c-ad06-40e4-905b-981c2f7cf6d4","Type":"ContainerStarted","Data":"20e179235b9b77103ae0d02d86fa5f59e56d9118d17fe0cb90bfe30cda4c44b9"} Oct 02 16:45:02 crc kubenswrapper[4808]: I1002 16:45:02.694127 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd" event={"ID":"260610f2-c80c-43f5-af6b-b4a4352a1564","Type":"ContainerStarted","Data":"86d0bc98fcb7b41348ef803c0dcd5ec9dda39bcf5ae56c116299049185485f6e"} Oct 02 16:45:02 crc kubenswrapper[4808]: I1002 16:45:02.696605 4808 generic.go:334] "Generic (PLEG): container finished" podID="b7ae49e8-10fa-4144-aa62-30d63a77ab0e" containerID="b0db62942881e3c571aa796de4d78cb5a3d6eff259ae174944bb069b945d2d77" exitCode=0 Oct 02 16:45:02 crc kubenswrapper[4808]: I1002 16:45:02.696660 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4q65f" event={"ID":"b7ae49e8-10fa-4144-aa62-30d63a77ab0e","Type":"ContainerDied","Data":"b0db62942881e3c571aa796de4d78cb5a3d6eff259ae174944bb069b945d2d77"} Oct 02 16:45:02 crc kubenswrapper[4808]: I1002 16:45:02.718211 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n6gnm" podStartSLOduration=3.162937725 podStartE2EDuration="5.718184005s" podCreationTimestamp="2025-10-02 16:44:57 +0000 UTC" firstStartedPulling="2025-10-02 16:44:58.653419531 +0000 UTC m=+265.978948531" lastFinishedPulling="2025-10-02 16:45:01.208665801 +0000 UTC m=+268.534194811" observedRunningTime="2025-10-02 16:45:02.712464061 +0000 UTC m=+270.037993071" watchObservedRunningTime="2025-10-02 16:45:02.718184005 +0000 UTC m=+270.043713005" Oct 02 16:45:02 crc kubenswrapper[4808]: I1002 16:45:02.734185 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd" podStartSLOduration=2.734161008 podStartE2EDuration="2.734161008s" podCreationTimestamp="2025-10-02 16:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:45:02.732177177 +0000 UTC m=+270.057706177" watchObservedRunningTime="2025-10-02 16:45:02.734161008 +0000 UTC m=+270.059690008" Oct 02 16:45:03 crc kubenswrapper[4808]: I1002 16:45:03.704477 4808 generic.go:334] "Generic (PLEG): container finished" podID="260610f2-c80c-43f5-af6b-b4a4352a1564" containerID="86d0bc98fcb7b41348ef803c0dcd5ec9dda39bcf5ae56c116299049185485f6e" exitCode=0 Oct 02 16:45:03 crc kubenswrapper[4808]: I1002 16:45:03.704597 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd" event={"ID":"260610f2-c80c-43f5-af6b-b4a4352a1564","Type":"ContainerDied","Data":"86d0bc98fcb7b41348ef803c0dcd5ec9dda39bcf5ae56c116299049185485f6e"} Oct 02 16:45:03 crc kubenswrapper[4808]: I1002 16:45:03.709677 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4q65f" event={"ID":"b7ae49e8-10fa-4144-aa62-30d63a77ab0e","Type":"ContainerStarted","Data":"8d65058c9af056a8d2bb0d7187ae5d92ad0caeadafa3afd01de5037124fd22b4"} Oct 02 16:45:04 crc kubenswrapper[4808]: I1002 16:45:04.993846 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd" Oct 02 16:45:05 crc kubenswrapper[4808]: I1002 16:45:05.013070 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4q65f" podStartSLOduration=4.504348267 podStartE2EDuration="8.013047521s" podCreationTimestamp="2025-10-02 16:44:57 +0000 UTC" firstStartedPulling="2025-10-02 16:44:59.663604173 +0000 UTC m=+266.989133173" lastFinishedPulling="2025-10-02 16:45:03.172303427 +0000 UTC m=+270.497832427" observedRunningTime="2025-10-02 16:45:03.740844507 +0000 UTC m=+271.066373507" watchObservedRunningTime="2025-10-02 16:45:05.013047521 +0000 UTC m=+272.338576521" Oct 02 16:45:05 crc kubenswrapper[4808]: I1002 16:45:05.120986 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nbjkh" Oct 02 16:45:05 crc kubenswrapper[4808]: I1002 16:45:05.121433 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nbjkh" Oct 02 16:45:05 crc kubenswrapper[4808]: I1002 16:45:05.177118 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nbjkh" Oct 02 16:45:05 crc kubenswrapper[4808]: I1002 16:45:05.191548 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/260610f2-c80c-43f5-af6b-b4a4352a1564-secret-volume\") pod \"260610f2-c80c-43f5-af6b-b4a4352a1564\" (UID: \"260610f2-c80c-43f5-af6b-b4a4352a1564\") " Oct 02 16:45:05 crc kubenswrapper[4808]: I1002 16:45:05.191701 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ks2hp\" (UniqueName: \"kubernetes.io/projected/260610f2-c80c-43f5-af6b-b4a4352a1564-kube-api-access-ks2hp\") pod \"260610f2-c80c-43f5-af6b-b4a4352a1564\" (UID: \"260610f2-c80c-43f5-af6b-b4a4352a1564\") " Oct 02 16:45:05 crc kubenswrapper[4808]: I1002 16:45:05.191752 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/260610f2-c80c-43f5-af6b-b4a4352a1564-config-volume\") pod \"260610f2-c80c-43f5-af6b-b4a4352a1564\" (UID: \"260610f2-c80c-43f5-af6b-b4a4352a1564\") " Oct 02 16:45:05 crc kubenswrapper[4808]: I1002 16:45:05.193294 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/260610f2-c80c-43f5-af6b-b4a4352a1564-config-volume" (OuterVolumeSpecName: "config-volume") pod "260610f2-c80c-43f5-af6b-b4a4352a1564" (UID: "260610f2-c80c-43f5-af6b-b4a4352a1564"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:45:05 crc kubenswrapper[4808]: I1002 16:45:05.202377 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/260610f2-c80c-43f5-af6b-b4a4352a1564-kube-api-access-ks2hp" (OuterVolumeSpecName: "kube-api-access-ks2hp") pod "260610f2-c80c-43f5-af6b-b4a4352a1564" (UID: "260610f2-c80c-43f5-af6b-b4a4352a1564"). InnerVolumeSpecName "kube-api-access-ks2hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:45:05 crc kubenswrapper[4808]: I1002 16:45:05.204051 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/260610f2-c80c-43f5-af6b-b4a4352a1564-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "260610f2-c80c-43f5-af6b-b4a4352a1564" (UID: "260610f2-c80c-43f5-af6b-b4a4352a1564"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:45:05 crc kubenswrapper[4808]: I1002 16:45:05.294096 4808 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/260610f2-c80c-43f5-af6b-b4a4352a1564-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 16:45:05 crc kubenswrapper[4808]: I1002 16:45:05.294152 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ks2hp\" (UniqueName: \"kubernetes.io/projected/260610f2-c80c-43f5-af6b-b4a4352a1564-kube-api-access-ks2hp\") on node \"crc\" DevicePath \"\"" Oct 02 16:45:05 crc kubenswrapper[4808]: I1002 16:45:05.294168 4808 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/260610f2-c80c-43f5-af6b-b4a4352a1564-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 16:45:05 crc kubenswrapper[4808]: I1002 16:45:05.723614 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd" event={"ID":"260610f2-c80c-43f5-af6b-b4a4352a1564","Type":"ContainerDied","Data":"06b3ae2aabb32dc1e7a874e250040a26be3c49061a174e9333c2ee6f30de37d6"} Oct 02 16:45:05 crc kubenswrapper[4808]: I1002 16:45:05.723687 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06b3ae2aabb32dc1e7a874e250040a26be3c49061a174e9333c2ee6f30de37d6" Oct 02 16:45:05 crc kubenswrapper[4808]: I1002 16:45:05.723637 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323725-nz4kd" Oct 02 16:45:05 crc kubenswrapper[4808]: I1002 16:45:05.780949 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nbjkh" Oct 02 16:45:05 crc kubenswrapper[4808]: I1002 16:45:05.787551 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k8j2q" Oct 02 16:45:05 crc kubenswrapper[4808]: I1002 16:45:05.787608 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k8j2q" Oct 02 16:45:05 crc kubenswrapper[4808]: I1002 16:45:05.852419 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k8j2q" Oct 02 16:45:06 crc kubenswrapper[4808]: I1002 16:45:06.779433 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k8j2q" Oct 02 16:45:07 crc kubenswrapper[4808]: I1002 16:45:07.529611 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n6gnm" Oct 02 16:45:07 crc kubenswrapper[4808]: I1002 16:45:07.530096 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n6gnm" Oct 02 16:45:07 crc kubenswrapper[4808]: I1002 16:45:07.573594 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n6gnm" Oct 02 16:45:07 crc kubenswrapper[4808]: I1002 16:45:07.778042 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n6gnm" Oct 02 16:45:08 crc kubenswrapper[4808]: I1002 16:45:08.115891 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4q65f" Oct 02 16:45:08 crc kubenswrapper[4808]: I1002 16:45:08.116524 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4q65f" Oct 02 16:45:08 crc kubenswrapper[4808]: I1002 16:45:08.167286 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4q65f" Oct 02 16:45:08 crc kubenswrapper[4808]: I1002 16:45:08.792466 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4q65f" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.283346 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-8974h" podUID="8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" containerName="oauth-openshift" containerID="cri-o://509f66c4aee579b514a27a4db21816ff20c9e09aaf37c319cbb4a7addffac342" gracePeriod=15 Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.764191 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.802499 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7865b47677-rlpzk"] Oct 02 16:45:19 crc kubenswrapper[4808]: E1002 16:45:19.803071 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="260610f2-c80c-43f5-af6b-b4a4352a1564" containerName="collect-profiles" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.803218 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="260610f2-c80c-43f5-af6b-b4a4352a1564" containerName="collect-profiles" Oct 02 16:45:19 crc kubenswrapper[4808]: E1002 16:45:19.803337 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" containerName="oauth-openshift" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.803466 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" containerName="oauth-openshift" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.803712 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="260610f2-c80c-43f5-af6b-b4a4352a1564" containerName="collect-profiles" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.803830 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" containerName="oauth-openshift" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.804562 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.816729 4808 generic.go:334] "Generic (PLEG): container finished" podID="8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" containerID="509f66c4aee579b514a27a4db21816ff20c9e09aaf37c319cbb4a7addffac342" exitCode=0 Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.816787 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8974h" event={"ID":"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324","Type":"ContainerDied","Data":"509f66c4aee579b514a27a4db21816ff20c9e09aaf37c319cbb4a7addffac342"} Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.816829 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8974h" event={"ID":"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324","Type":"ContainerDied","Data":"82bce0aaceee20db1d78fbc7371d275ab6a03fcafeaab7dc7cd77d300798735d"} Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.816851 4808 scope.go:117] "RemoveContainer" containerID="509f66c4aee579b514a27a4db21816ff20c9e09aaf37c319cbb4a7addffac342" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.816974 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8974h" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.821842 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7865b47677-rlpzk"] Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.825380 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.825417 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.825455 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.825518 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-session\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.825552 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.825575 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b7517656-9d2d-488c-9149-ce858ccf2e8b-audit-policies\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.825596 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.825614 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-service-ca\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.825641 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-user-template-login\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.825664 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-user-template-error\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.825684 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.825707 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b7517656-9d2d-488c-9149-ce858ccf2e8b-audit-dir\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.825725 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgdsd\" (UniqueName: \"kubernetes.io/projected/b7517656-9d2d-488c-9149-ce858ccf2e8b-kube-api-access-lgdsd\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.825746 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-router-certs\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.841592 4808 scope.go:117] "RemoveContainer" containerID="509f66c4aee579b514a27a4db21816ff20c9e09aaf37c319cbb4a7addffac342" Oct 02 16:45:19 crc kubenswrapper[4808]: E1002 16:45:19.842338 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"509f66c4aee579b514a27a4db21816ff20c9e09aaf37c319cbb4a7addffac342\": container with ID starting with 509f66c4aee579b514a27a4db21816ff20c9e09aaf37c319cbb4a7addffac342 not found: ID does not exist" containerID="509f66c4aee579b514a27a4db21816ff20c9e09aaf37c319cbb4a7addffac342" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.842404 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"509f66c4aee579b514a27a4db21816ff20c9e09aaf37c319cbb4a7addffac342"} err="failed to get container status \"509f66c4aee579b514a27a4db21816ff20c9e09aaf37c319cbb4a7addffac342\": rpc error: code = NotFound desc = could not find container \"509f66c4aee579b514a27a4db21816ff20c9e09aaf37c319cbb4a7addffac342\": container with ID starting with 509f66c4aee579b514a27a4db21816ff20c9e09aaf37c319cbb4a7addffac342 not found: ID does not exist" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.926658 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-template-provider-selection\") pod \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.927088 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-service-ca\") pod \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.928184 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-idp-0-file-data\") pod \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.928808 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-template-login\") pod \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.928954 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-ocp-branding-template\") pod \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.929099 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-router-certs\") pod \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.929284 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-serving-cert\") pod \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.929395 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-template-error\") pod \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.929521 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-audit-policies\") pod \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.928099 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" (UID: "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.929611 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-audit-dir\") pod \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.929803 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-cliconfig\") pod \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.929874 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkxxv\" (UniqueName: \"kubernetes.io/projected/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-kube-api-access-zkxxv\") pod \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.929901 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-trusted-ca-bundle\") pod \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.929898 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" (UID: "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.929959 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-session\") pod \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\" (UID: \"8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324\") " Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.930136 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" (UID: "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.930158 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" (UID: "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.930280 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b7517656-9d2d-488c-9149-ce858ccf2e8b-audit-dir\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.930391 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgdsd\" (UniqueName: \"kubernetes.io/projected/b7517656-9d2d-488c-9149-ce858ccf2e8b-kube-api-access-lgdsd\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.930494 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-router-certs\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.930614 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.930715 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.930324 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b7517656-9d2d-488c-9149-ce858ccf2e8b-audit-dir\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.930745 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" (UID: "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.939097 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" (UID: "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.939741 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" (UID: "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.940118 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" (UID: "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.931921 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.938223 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.942854 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.942215 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.943220 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-session\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.943426 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.943547 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b7517656-9d2d-488c-9149-ce858ccf2e8b-audit-policies\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.943625 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.943702 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-service-ca\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.943818 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-user-template-login\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.943899 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-user-template-error\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.943973 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.944104 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.944175 4808 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.944260 4808 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.944330 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.944393 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.944473 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.944554 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.944618 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.944269 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-router-certs\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.945190 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-service-ca\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.948538 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b7517656-9d2d-488c-9149-ce858ccf2e8b-audit-policies\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.950184 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-user-template-login\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.950716 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-user-template-error\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.951021 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-session\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.951599 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.953792 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" (UID: "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.954293 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" (UID: "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.954384 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-kube-api-access-zkxxv" (OuterVolumeSpecName: "kube-api-access-zkxxv") pod "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" (UID: "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324"). InnerVolumeSpecName "kube-api-access-zkxxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.954793 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.955186 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b7517656-9d2d-488c-9149-ce858ccf2e8b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.955719 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgdsd\" (UniqueName: \"kubernetes.io/projected/b7517656-9d2d-488c-9149-ce858ccf2e8b-kube-api-access-lgdsd\") pod \"oauth-openshift-7865b47677-rlpzk\" (UID: \"b7517656-9d2d-488c-9149-ce858ccf2e8b\") " pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.963093 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" (UID: "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.963881 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" (UID: "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:45:19 crc kubenswrapper[4808]: I1002 16:45:19.964313 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" (UID: "8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:45:20 crc kubenswrapper[4808]: I1002 16:45:20.046106 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:45:20 crc kubenswrapper[4808]: I1002 16:45:20.046408 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 02 16:45:20 crc kubenswrapper[4808]: I1002 16:45:20.046519 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:45:20 crc kubenswrapper[4808]: I1002 16:45:20.046595 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkxxv\" (UniqueName: \"kubernetes.io/projected/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-kube-api-access-zkxxv\") on node \"crc\" DevicePath \"\"" Oct 02 16:45:20 crc kubenswrapper[4808]: I1002 16:45:20.046668 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 02 16:45:20 crc kubenswrapper[4808]: I1002 16:45:20.046739 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 02 16:45:20 crc kubenswrapper[4808]: I1002 16:45:20.123081 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:20 crc kubenswrapper[4808]: I1002 16:45:20.156664 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8974h"] Oct 02 16:45:20 crc kubenswrapper[4808]: I1002 16:45:20.165275 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8974h"] Oct 02 16:45:20 crc kubenswrapper[4808]: I1002 16:45:20.559815 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7865b47677-rlpzk"] Oct 02 16:45:20 crc kubenswrapper[4808]: W1002 16:45:20.564738 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7517656_9d2d_488c_9149_ce858ccf2e8b.slice/crio-7c14769077cefb4d77a8aeef1e2ca4309abccf44455f459637c0280c0ed1ce88 WatchSource:0}: Error finding container 7c14769077cefb4d77a8aeef1e2ca4309abccf44455f459637c0280c0ed1ce88: Status 404 returned error can't find the container with id 7c14769077cefb4d77a8aeef1e2ca4309abccf44455f459637c0280c0ed1ce88 Oct 02 16:45:20 crc kubenswrapper[4808]: I1002 16:45:20.826307 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" event={"ID":"b7517656-9d2d-488c-9149-ce858ccf2e8b","Type":"ContainerStarted","Data":"7c14769077cefb4d77a8aeef1e2ca4309abccf44455f459637c0280c0ed1ce88"} Oct 02 16:45:21 crc kubenswrapper[4808]: I1002 16:45:21.403336 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324" path="/var/lib/kubelet/pods/8c9e6c9c-11b2-410c-ab6c-e6b8c25aa324/volumes" Oct 02 16:45:21 crc kubenswrapper[4808]: I1002 16:45:21.835405 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" event={"ID":"b7517656-9d2d-488c-9149-ce858ccf2e8b","Type":"ContainerStarted","Data":"bb8188680757d713c894c663086eb3a68555344a2994b380fac7286afb57a097"} Oct 02 16:45:21 crc kubenswrapper[4808]: I1002 16:45:21.835773 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:21 crc kubenswrapper[4808]: I1002 16:45:21.841428 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" Oct 02 16:45:21 crc kubenswrapper[4808]: I1002 16:45:21.871000 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7865b47677-rlpzk" podStartSLOduration=27.87097232 podStartE2EDuration="27.87097232s" podCreationTimestamp="2025-10-02 16:44:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:45:21.870001393 +0000 UTC m=+289.195530413" watchObservedRunningTime="2025-10-02 16:45:21.87097232 +0000 UTC m=+289.196501320" Oct 02 16:46:48 crc kubenswrapper[4808]: I1002 16:46:48.739444 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 16:46:48 crc kubenswrapper[4808]: I1002 16:46:48.740015 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 16:47:18 crc kubenswrapper[4808]: I1002 16:47:18.739637 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 16:47:18 crc kubenswrapper[4808]: I1002 16:47:18.740556 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 16:47:48 crc kubenswrapper[4808]: I1002 16:47:48.740035 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 16:47:48 crc kubenswrapper[4808]: I1002 16:47:48.742830 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 16:47:48 crc kubenswrapper[4808]: I1002 16:47:48.743133 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 16:47:48 crc kubenswrapper[4808]: I1002 16:47:48.744510 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f801630a113a03b3e035f8940f01d306b49a7227476e62d1552f3229787c1924"} pod="openshift-machine-config-operator/machine-config-daemon-7z66r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 16:47:48 crc kubenswrapper[4808]: I1002 16:47:48.744797 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" containerID="cri-o://f801630a113a03b3e035f8940f01d306b49a7227476e62d1552f3229787c1924" gracePeriod=600 Oct 02 16:47:49 crc kubenswrapper[4808]: I1002 16:47:49.878080 4808 generic.go:334] "Generic (PLEG): container finished" podID="f83ce425-101d-4489-94a4-5c256eb29328" containerID="f801630a113a03b3e035f8940f01d306b49a7227476e62d1552f3229787c1924" exitCode=0 Oct 02 16:47:49 crc kubenswrapper[4808]: I1002 16:47:49.878196 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" event={"ID":"f83ce425-101d-4489-94a4-5c256eb29328","Type":"ContainerDied","Data":"f801630a113a03b3e035f8940f01d306b49a7227476e62d1552f3229787c1924"} Oct 02 16:47:49 crc kubenswrapper[4808]: I1002 16:47:49.879762 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" event={"ID":"f83ce425-101d-4489-94a4-5c256eb29328","Type":"ContainerStarted","Data":"a9406662e09c7be4b59f910ebd57fef50f5b64e0992e5ddee0ca92a0e3d8521a"} Oct 02 16:47:49 crc kubenswrapper[4808]: I1002 16:47:49.879900 4808 scope.go:117] "RemoveContainer" containerID="4cf36121a19f29c215024c781fe05ece8def249a16ae0e496c2660cbc423cfba" Oct 02 16:48:19 crc kubenswrapper[4808]: I1002 16:48:19.983796 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rjgkx"] Oct 02 16:48:19 crc kubenswrapper[4808]: I1002 16:48:19.985787 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.008265 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rjgkx"] Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.096080 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-bound-sa-token\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.096144 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-trusted-ca\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.096180 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.096311 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-registry-certificates\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.096385 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggw5b\" (UniqueName: \"kubernetes.io/projected/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-kube-api-access-ggw5b\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.096436 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.096477 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.096565 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-registry-tls\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.125046 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.198356 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-bound-sa-token\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.198399 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-trusted-ca\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.198425 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.198441 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-registry-certificates\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.198459 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggw5b\" (UniqueName: \"kubernetes.io/projected/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-kube-api-access-ggw5b\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.198492 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.198516 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-registry-tls\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.199352 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.200161 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-registry-certificates\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.200741 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-trusted-ca\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.204187 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-registry-tls\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.211943 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.218816 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggw5b\" (UniqueName: \"kubernetes.io/projected/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-kube-api-access-ggw5b\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.219485 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/21afeeaf-cd4b-49f2-a0b8-e3173b90e351-bound-sa-token\") pod \"image-registry-66df7c8f76-rjgkx\" (UID: \"21afeeaf-cd4b-49f2-a0b8-e3173b90e351\") " pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.309497 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:20 crc kubenswrapper[4808]: I1002 16:48:20.576898 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rjgkx"] Oct 02 16:48:20 crc kubenswrapper[4808]: W1002 16:48:20.588786 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21afeeaf_cd4b_49f2_a0b8_e3173b90e351.slice/crio-7b1573a1efa254d5b694d4bbf32ea99d7dd18a535197b769671f0035fe39cacf WatchSource:0}: Error finding container 7b1573a1efa254d5b694d4bbf32ea99d7dd18a535197b769671f0035fe39cacf: Status 404 returned error can't find the container with id 7b1573a1efa254d5b694d4bbf32ea99d7dd18a535197b769671f0035fe39cacf Oct 02 16:48:21 crc kubenswrapper[4808]: I1002 16:48:21.127318 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" event={"ID":"21afeeaf-cd4b-49f2-a0b8-e3173b90e351","Type":"ContainerStarted","Data":"1b5c7ad678f6cf341b4d6ae0e8688d5b7526356fd79d021d18077a14ac656d3f"} Oct 02 16:48:21 crc kubenswrapper[4808]: I1002 16:48:21.127393 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" event={"ID":"21afeeaf-cd4b-49f2-a0b8-e3173b90e351","Type":"ContainerStarted","Data":"7b1573a1efa254d5b694d4bbf32ea99d7dd18a535197b769671f0035fe39cacf"} Oct 02 16:48:21 crc kubenswrapper[4808]: I1002 16:48:21.128468 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:21 crc kubenswrapper[4808]: I1002 16:48:21.152570 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" podStartSLOduration=2.152539093 podStartE2EDuration="2.152539093s" podCreationTimestamp="2025-10-02 16:48:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:48:21.149082691 +0000 UTC m=+468.474611691" watchObservedRunningTime="2025-10-02 16:48:21.152539093 +0000 UTC m=+468.478068123" Oct 02 16:48:40 crc kubenswrapper[4808]: I1002 16:48:40.316541 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-rjgkx" Oct 02 16:48:40 crc kubenswrapper[4808]: I1002 16:48:40.386990 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9vv2b"] Oct 02 16:49:05 crc kubenswrapper[4808]: I1002 16:49:05.449570 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" podUID="71aafdd1-5917-4764-9503-4c266977f251" containerName="registry" containerID="cri-o://acc25e658141f487455f0b3f81462b05c303df4035f8ec9898137ee3282935fc" gracePeriod=30 Oct 02 16:49:05 crc kubenswrapper[4808]: I1002 16:49:05.940622 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.124104 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71aafdd1-5917-4764-9503-4c266977f251-trusted-ca\") pod \"71aafdd1-5917-4764-9503-4c266977f251\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.124291 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/71aafdd1-5917-4764-9503-4c266977f251-registry-certificates\") pod \"71aafdd1-5917-4764-9503-4c266977f251\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.126006 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71aafdd1-5917-4764-9503-4c266977f251-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "71aafdd1-5917-4764-9503-4c266977f251" (UID: "71aafdd1-5917-4764-9503-4c266977f251"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.126135 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zn6kd\" (UniqueName: \"kubernetes.io/projected/71aafdd1-5917-4764-9503-4c266977f251-kube-api-access-zn6kd\") pod \"71aafdd1-5917-4764-9503-4c266977f251\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.126530 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71aafdd1-5917-4764-9503-4c266977f251-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "71aafdd1-5917-4764-9503-4c266977f251" (UID: "71aafdd1-5917-4764-9503-4c266977f251"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.127479 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"71aafdd1-5917-4764-9503-4c266977f251\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.127860 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/71aafdd1-5917-4764-9503-4c266977f251-registry-tls\") pod \"71aafdd1-5917-4764-9503-4c266977f251\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.127937 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/71aafdd1-5917-4764-9503-4c266977f251-ca-trust-extracted\") pod \"71aafdd1-5917-4764-9503-4c266977f251\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.127982 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/71aafdd1-5917-4764-9503-4c266977f251-installation-pull-secrets\") pod \"71aafdd1-5917-4764-9503-4c266977f251\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.128014 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/71aafdd1-5917-4764-9503-4c266977f251-bound-sa-token\") pod \"71aafdd1-5917-4764-9503-4c266977f251\" (UID: \"71aafdd1-5917-4764-9503-4c266977f251\") " Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.128842 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71aafdd1-5917-4764-9503-4c266977f251-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.128882 4808 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/71aafdd1-5917-4764-9503-4c266977f251-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.139079 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71aafdd1-5917-4764-9503-4c266977f251-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "71aafdd1-5917-4764-9503-4c266977f251" (UID: "71aafdd1-5917-4764-9503-4c266977f251"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.139310 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71aafdd1-5917-4764-9503-4c266977f251-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "71aafdd1-5917-4764-9503-4c266977f251" (UID: "71aafdd1-5917-4764-9503-4c266977f251"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.139841 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71aafdd1-5917-4764-9503-4c266977f251-kube-api-access-zn6kd" (OuterVolumeSpecName: "kube-api-access-zn6kd") pod "71aafdd1-5917-4764-9503-4c266977f251" (UID: "71aafdd1-5917-4764-9503-4c266977f251"). InnerVolumeSpecName "kube-api-access-zn6kd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.143678 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71aafdd1-5917-4764-9503-4c266977f251-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "71aafdd1-5917-4764-9503-4c266977f251" (UID: "71aafdd1-5917-4764-9503-4c266977f251"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.144713 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "71aafdd1-5917-4764-9503-4c266977f251" (UID: "71aafdd1-5917-4764-9503-4c266977f251"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.163739 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71aafdd1-5917-4764-9503-4c266977f251-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "71aafdd1-5917-4764-9503-4c266977f251" (UID: "71aafdd1-5917-4764-9503-4c266977f251"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.229738 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zn6kd\" (UniqueName: \"kubernetes.io/projected/71aafdd1-5917-4764-9503-4c266977f251-kube-api-access-zn6kd\") on node \"crc\" DevicePath \"\"" Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.229783 4808 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/71aafdd1-5917-4764-9503-4c266977f251-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.229796 4808 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/71aafdd1-5917-4764-9503-4c266977f251-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.229808 4808 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/71aafdd1-5917-4764-9503-4c266977f251-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.229823 4808 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/71aafdd1-5917-4764-9503-4c266977f251-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.446308 4808 generic.go:334] "Generic (PLEG): container finished" podID="71aafdd1-5917-4764-9503-4c266977f251" containerID="acc25e658141f487455f0b3f81462b05c303df4035f8ec9898137ee3282935fc" exitCode=0 Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.446354 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" event={"ID":"71aafdd1-5917-4764-9503-4c266977f251","Type":"ContainerDied","Data":"acc25e658141f487455f0b3f81462b05c303df4035f8ec9898137ee3282935fc"} Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.446389 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" event={"ID":"71aafdd1-5917-4764-9503-4c266977f251","Type":"ContainerDied","Data":"4387e5f3b5a1f67e79bf621dffeacc50959b134dd741e2a4102f50d751c048b7"} Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.446418 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9vv2b" Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.446430 4808 scope.go:117] "RemoveContainer" containerID="acc25e658141f487455f0b3f81462b05c303df4035f8ec9898137ee3282935fc" Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.483355 4808 scope.go:117] "RemoveContainer" containerID="acc25e658141f487455f0b3f81462b05c303df4035f8ec9898137ee3282935fc" Oct 02 16:49:06 crc kubenswrapper[4808]: E1002 16:49:06.483990 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acc25e658141f487455f0b3f81462b05c303df4035f8ec9898137ee3282935fc\": container with ID starting with acc25e658141f487455f0b3f81462b05c303df4035f8ec9898137ee3282935fc not found: ID does not exist" containerID="acc25e658141f487455f0b3f81462b05c303df4035f8ec9898137ee3282935fc" Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.484048 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acc25e658141f487455f0b3f81462b05c303df4035f8ec9898137ee3282935fc"} err="failed to get container status \"acc25e658141f487455f0b3f81462b05c303df4035f8ec9898137ee3282935fc\": rpc error: code = NotFound desc = could not find container \"acc25e658141f487455f0b3f81462b05c303df4035f8ec9898137ee3282935fc\": container with ID starting with acc25e658141f487455f0b3f81462b05c303df4035f8ec9898137ee3282935fc not found: ID does not exist" Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.493812 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9vv2b"] Oct 02 16:49:06 crc kubenswrapper[4808]: I1002 16:49:06.497364 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9vv2b"] Oct 02 16:49:07 crc kubenswrapper[4808]: I1002 16:49:07.411026 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71aafdd1-5917-4764-9503-4c266977f251" path="/var/lib/kubelet/pods/71aafdd1-5917-4764-9503-4c266977f251/volumes" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.373325 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-fwdv4"] Oct 02 16:50:12 crc kubenswrapper[4808]: E1002 16:50:12.374136 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71aafdd1-5917-4764-9503-4c266977f251" containerName="registry" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.374154 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="71aafdd1-5917-4764-9503-4c266977f251" containerName="registry" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.374353 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="71aafdd1-5917-4764-9503-4c266977f251" containerName="registry" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.374866 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-fwdv4" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.376706 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-5s7x8"] Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.377504 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-5s7x8" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.377639 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.378089 4808 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-jn4fc" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.378297 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.380301 4808 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-sc8r9" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.381798 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-5s7x8"] Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.391342 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-fwdv4"] Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.418830 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-gp2gc"] Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.419641 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-gp2gc" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.435927 4808 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-zt8pm" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.441477 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-gp2gc"] Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.454308 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frd7w\" (UniqueName: \"kubernetes.io/projected/dfc994a0-86f0-4358-a2f3-ff474b5e48d6-kube-api-access-frd7w\") pod \"cert-manager-cainjector-7f985d654d-5s7x8\" (UID: \"dfc994a0-86f0-4358-a2f3-ff474b5e48d6\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-5s7x8" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.555857 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frd7w\" (UniqueName: \"kubernetes.io/projected/dfc994a0-86f0-4358-a2f3-ff474b5e48d6-kube-api-access-frd7w\") pod \"cert-manager-cainjector-7f985d654d-5s7x8\" (UID: \"dfc994a0-86f0-4358-a2f3-ff474b5e48d6\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-5s7x8" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.555978 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn4sv\" (UniqueName: \"kubernetes.io/projected/afc013f6-182d-4d68-bb21-7e933464712b-kube-api-access-jn4sv\") pod \"cert-manager-webhook-5655c58dd6-gp2gc\" (UID: \"afc013f6-182d-4d68-bb21-7e933464712b\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-gp2gc" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.556011 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbvj4\" (UniqueName: \"kubernetes.io/projected/d45bb58b-f832-4f1c-92d0-286ba9ca7ab9-kube-api-access-pbvj4\") pod \"cert-manager-5b446d88c5-fwdv4\" (UID: \"d45bb58b-f832-4f1c-92d0-286ba9ca7ab9\") " pod="cert-manager/cert-manager-5b446d88c5-fwdv4" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.575301 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frd7w\" (UniqueName: \"kubernetes.io/projected/dfc994a0-86f0-4358-a2f3-ff474b5e48d6-kube-api-access-frd7w\") pod \"cert-manager-cainjector-7f985d654d-5s7x8\" (UID: \"dfc994a0-86f0-4358-a2f3-ff474b5e48d6\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-5s7x8" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.657271 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn4sv\" (UniqueName: \"kubernetes.io/projected/afc013f6-182d-4d68-bb21-7e933464712b-kube-api-access-jn4sv\") pod \"cert-manager-webhook-5655c58dd6-gp2gc\" (UID: \"afc013f6-182d-4d68-bb21-7e933464712b\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-gp2gc" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.657627 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbvj4\" (UniqueName: \"kubernetes.io/projected/d45bb58b-f832-4f1c-92d0-286ba9ca7ab9-kube-api-access-pbvj4\") pod \"cert-manager-5b446d88c5-fwdv4\" (UID: \"d45bb58b-f832-4f1c-92d0-286ba9ca7ab9\") " pod="cert-manager/cert-manager-5b446d88c5-fwdv4" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.682879 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbvj4\" (UniqueName: \"kubernetes.io/projected/d45bb58b-f832-4f1c-92d0-286ba9ca7ab9-kube-api-access-pbvj4\") pod \"cert-manager-5b446d88c5-fwdv4\" (UID: \"d45bb58b-f832-4f1c-92d0-286ba9ca7ab9\") " pod="cert-manager/cert-manager-5b446d88c5-fwdv4" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.689204 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn4sv\" (UniqueName: \"kubernetes.io/projected/afc013f6-182d-4d68-bb21-7e933464712b-kube-api-access-jn4sv\") pod \"cert-manager-webhook-5655c58dd6-gp2gc\" (UID: \"afc013f6-182d-4d68-bb21-7e933464712b\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-gp2gc" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.697165 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-fwdv4" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.710848 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-5s7x8" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.743943 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-gp2gc" Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.936278 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-fwdv4"] Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.947162 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 16:50:12 crc kubenswrapper[4808]: I1002 16:50:12.983514 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-5s7x8"] Oct 02 16:50:12 crc kubenswrapper[4808]: W1002 16:50:12.986262 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddfc994a0_86f0_4358_a2f3_ff474b5e48d6.slice/crio-4e720d038264a3b0a4e04ee6c23470594412b2954a93af783af4babf73cfaf8a WatchSource:0}: Error finding container 4e720d038264a3b0a4e04ee6c23470594412b2954a93af783af4babf73cfaf8a: Status 404 returned error can't find the container with id 4e720d038264a3b0a4e04ee6c23470594412b2954a93af783af4babf73cfaf8a Oct 02 16:50:13 crc kubenswrapper[4808]: I1002 16:50:13.264201 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-gp2gc"] Oct 02 16:50:13 crc kubenswrapper[4808]: W1002 16:50:13.268073 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafc013f6_182d_4d68_bb21_7e933464712b.slice/crio-0cda77eb3ef4f7707a38c80248410ad4f7a8153365c5aebf995f002e3d86e80d WatchSource:0}: Error finding container 0cda77eb3ef4f7707a38c80248410ad4f7a8153365c5aebf995f002e3d86e80d: Status 404 returned error can't find the container with id 0cda77eb3ef4f7707a38c80248410ad4f7a8153365c5aebf995f002e3d86e80d Oct 02 16:50:13 crc kubenswrapper[4808]: I1002 16:50:13.911733 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-fwdv4" event={"ID":"d45bb58b-f832-4f1c-92d0-286ba9ca7ab9","Type":"ContainerStarted","Data":"0471e85c75a315befd22c49df8a2f456b64d0bc9cf35d905b4bd91ee1d148181"} Oct 02 16:50:13 crc kubenswrapper[4808]: I1002 16:50:13.912777 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-5s7x8" event={"ID":"dfc994a0-86f0-4358-a2f3-ff474b5e48d6","Type":"ContainerStarted","Data":"4e720d038264a3b0a4e04ee6c23470594412b2954a93af783af4babf73cfaf8a"} Oct 02 16:50:13 crc kubenswrapper[4808]: I1002 16:50:13.913635 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-gp2gc" event={"ID":"afc013f6-182d-4d68-bb21-7e933464712b","Type":"ContainerStarted","Data":"0cda77eb3ef4f7707a38c80248410ad4f7a8153365c5aebf995f002e3d86e80d"} Oct 02 16:50:16 crc kubenswrapper[4808]: I1002 16:50:16.943602 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-fwdv4" event={"ID":"d45bb58b-f832-4f1c-92d0-286ba9ca7ab9","Type":"ContainerStarted","Data":"945a9f9777bd20b18a7bff94456cac743eb4b448dac544e7eaba5ab6358cb103"} Oct 02 16:50:16 crc kubenswrapper[4808]: I1002 16:50:16.946564 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-5s7x8" event={"ID":"dfc994a0-86f0-4358-a2f3-ff474b5e48d6","Type":"ContainerStarted","Data":"06cf39f52bf50b81cc74d2ff1695328991627e1e32d520a0a7ad70b64e330608"} Oct 02 16:50:16 crc kubenswrapper[4808]: I1002 16:50:16.948195 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-gp2gc" event={"ID":"afc013f6-182d-4d68-bb21-7e933464712b","Type":"ContainerStarted","Data":"f1b3cb20e0ae1a19c2570e2347721fb5da577fea50eba214862be900a291a536"} Oct 02 16:50:16 crc kubenswrapper[4808]: I1002 16:50:16.948643 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-gp2gc" Oct 02 16:50:16 crc kubenswrapper[4808]: I1002 16:50:16.964893 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-fwdv4" podStartSLOduration=1.316333223 podStartE2EDuration="4.964874975s" podCreationTimestamp="2025-10-02 16:50:12 +0000 UTC" firstStartedPulling="2025-10-02 16:50:12.946974481 +0000 UTC m=+580.272503481" lastFinishedPulling="2025-10-02 16:50:16.595516193 +0000 UTC m=+583.921045233" observedRunningTime="2025-10-02 16:50:16.963651162 +0000 UTC m=+584.289180182" watchObservedRunningTime="2025-10-02 16:50:16.964874975 +0000 UTC m=+584.290403985" Oct 02 16:50:16 crc kubenswrapper[4808]: I1002 16:50:16.980648 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-gp2gc" podStartSLOduration=1.63965044 podStartE2EDuration="4.980606137s" podCreationTimestamp="2025-10-02 16:50:12 +0000 UTC" firstStartedPulling="2025-10-02 16:50:13.27070838 +0000 UTC m=+580.596237370" lastFinishedPulling="2025-10-02 16:50:16.611664057 +0000 UTC m=+583.937193067" observedRunningTime="2025-10-02 16:50:16.976383731 +0000 UTC m=+584.301912731" watchObservedRunningTime="2025-10-02 16:50:16.980606137 +0000 UTC m=+584.306135137" Oct 02 16:50:16 crc kubenswrapper[4808]: I1002 16:50:16.997585 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-5s7x8" podStartSLOduration=1.409690436 podStartE2EDuration="4.997566383s" podCreationTimestamp="2025-10-02 16:50:12 +0000 UTC" firstStartedPulling="2025-10-02 16:50:12.98845791 +0000 UTC m=+580.313986910" lastFinishedPulling="2025-10-02 16:50:16.576333817 +0000 UTC m=+583.901862857" observedRunningTime="2025-10-02 16:50:16.994937311 +0000 UTC m=+584.320466311" watchObservedRunningTime="2025-10-02 16:50:16.997566383 +0000 UTC m=+584.323095383" Oct 02 16:50:18 crc kubenswrapper[4808]: I1002 16:50:18.739910 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 16:50:18 crc kubenswrapper[4808]: I1002 16:50:18.739993 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 16:50:22 crc kubenswrapper[4808]: I1002 16:50:22.747222 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-gp2gc" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.362000 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zm46w"] Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.362530 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovn-controller" containerID="cri-o://2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f" gracePeriod=30 Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.362616 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="sbdb" containerID="cri-o://e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464" gracePeriod=30 Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.362714 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="nbdb" containerID="cri-o://15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0" gracePeriod=30 Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.362758 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="northd" containerID="cri-o://20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6" gracePeriod=30 Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.362792 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5" gracePeriod=30 Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.362829 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="kube-rbac-proxy-node" containerID="cri-o://964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa" gracePeriod=30 Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.362867 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovn-acl-logging" containerID="cri-o://827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200" gracePeriod=30 Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.427375 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovnkube-controller" containerID="cri-o://8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419" gracePeriod=30 Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.707456 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zm46w_e441e501-09ee-4e3d-a763-c853de09a02c/ovnkube-controller/3.log" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.711550 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zm46w_e441e501-09ee-4e3d-a763-c853de09a02c/ovn-acl-logging/0.log" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.712202 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zm46w_e441e501-09ee-4e3d-a763-c853de09a02c/ovn-controller/0.log" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.713052 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.790791 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-q6km5"] Oct 02 16:50:23 crc kubenswrapper[4808]: E1002 16:50:23.791151 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovnkube-controller" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791172 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovnkube-controller" Oct 02 16:50:23 crc kubenswrapper[4808]: E1002 16:50:23.791189 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovn-controller" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791205 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovn-controller" Oct 02 16:50:23 crc kubenswrapper[4808]: E1002 16:50:23.791226 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="sbdb" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791245 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="sbdb" Oct 02 16:50:23 crc kubenswrapper[4808]: E1002 16:50:23.791286 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovnkube-controller" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791298 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovnkube-controller" Oct 02 16:50:23 crc kubenswrapper[4808]: E1002 16:50:23.791317 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791329 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 16:50:23 crc kubenswrapper[4808]: E1002 16:50:23.791344 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="northd" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791356 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="northd" Oct 02 16:50:23 crc kubenswrapper[4808]: E1002 16:50:23.791402 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovnkube-controller" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791415 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovnkube-controller" Oct 02 16:50:23 crc kubenswrapper[4808]: E1002 16:50:23.791432 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="nbdb" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791444 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="nbdb" Oct 02 16:50:23 crc kubenswrapper[4808]: E1002 16:50:23.791462 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="kubecfg-setup" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791474 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="kubecfg-setup" Oct 02 16:50:23 crc kubenswrapper[4808]: E1002 16:50:23.791491 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="kube-rbac-proxy-node" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791504 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="kube-rbac-proxy-node" Oct 02 16:50:23 crc kubenswrapper[4808]: E1002 16:50:23.791523 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovn-acl-logging" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791538 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovn-acl-logging" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791702 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovnkube-controller" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791718 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovnkube-controller" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791733 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="sbdb" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791752 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="kube-rbac-proxy-node" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791770 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovn-acl-logging" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791789 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovnkube-controller" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791807 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791826 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="nbdb" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791843 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovn-controller" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.791859 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="northd" Oct 02 16:50:23 crc kubenswrapper[4808]: E1002 16:50:23.792009 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovnkube-controller" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.792024 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovnkube-controller" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.792210 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovnkube-controller" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.792227 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovnkube-controller" Oct 02 16:50:23 crc kubenswrapper[4808]: E1002 16:50:23.792446 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovnkube-controller" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.792461 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" containerName="ovnkube-controller" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.794977 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.808604 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"e441e501-09ee-4e3d-a763-c853de09a02c\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.808645 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-slash\") pod \"e441e501-09ee-4e3d-a763-c853de09a02c\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.808670 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-run-ovn-kubernetes\") pod \"e441e501-09ee-4e3d-a763-c853de09a02c\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.808695 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-cni-netd\") pod \"e441e501-09ee-4e3d-a763-c853de09a02c\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.808722 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwkmd\" (UniqueName: \"kubernetes.io/projected/e441e501-09ee-4e3d-a763-c853de09a02c-kube-api-access-bwkmd\") pod \"e441e501-09ee-4e3d-a763-c853de09a02c\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.808737 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-log-socket\") pod \"e441e501-09ee-4e3d-a763-c853de09a02c\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.808764 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-run-systemd\") pod \"e441e501-09ee-4e3d-a763-c853de09a02c\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.808761 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "e441e501-09ee-4e3d-a763-c853de09a02c" (UID: "e441e501-09ee-4e3d-a763-c853de09a02c"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.808787 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-cni-bin\") pod \"e441e501-09ee-4e3d-a763-c853de09a02c\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.808838 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "e441e501-09ee-4e3d-a763-c853de09a02c" (UID: "e441e501-09ee-4e3d-a763-c853de09a02c"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.808859 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-slash" (OuterVolumeSpecName: "host-slash") pod "e441e501-09ee-4e3d-a763-c853de09a02c" (UID: "e441e501-09ee-4e3d-a763-c853de09a02c"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.808905 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "e441e501-09ee-4e3d-a763-c853de09a02c" (UID: "e441e501-09ee-4e3d-a763-c853de09a02c"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.808903 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-log-socket" (OuterVolumeSpecName: "log-socket") pod "e441e501-09ee-4e3d-a763-c853de09a02c" (UID: "e441e501-09ee-4e3d-a763-c853de09a02c"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.808910 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "e441e501-09ee-4e3d-a763-c853de09a02c" (UID: "e441e501-09ee-4e3d-a763-c853de09a02c"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.808884 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-run-ovn\") pod \"e441e501-09ee-4e3d-a763-c853de09a02c\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809072 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-systemd-units\") pod \"e441e501-09ee-4e3d-a763-c853de09a02c\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809096 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-etc-openvswitch\") pod \"e441e501-09ee-4e3d-a763-c853de09a02c\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809157 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-node-log\") pod \"e441e501-09ee-4e3d-a763-c853de09a02c\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809184 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e441e501-09ee-4e3d-a763-c853de09a02c-ovn-node-metrics-cert\") pod \"e441e501-09ee-4e3d-a763-c853de09a02c\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809206 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-run-netns\") pod \"e441e501-09ee-4e3d-a763-c853de09a02c\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809281 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e441e501-09ee-4e3d-a763-c853de09a02c-ovnkube-script-lib\") pod \"e441e501-09ee-4e3d-a763-c853de09a02c\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809325 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e441e501-09ee-4e3d-a763-c853de09a02c-env-overrides\") pod \"e441e501-09ee-4e3d-a763-c853de09a02c\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809348 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-run-openvswitch\") pod \"e441e501-09ee-4e3d-a763-c853de09a02c\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809369 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e441e501-09ee-4e3d-a763-c853de09a02c-ovnkube-config\") pod \"e441e501-09ee-4e3d-a763-c853de09a02c\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809398 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-var-lib-openvswitch\") pod \"e441e501-09ee-4e3d-a763-c853de09a02c\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809430 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-kubelet\") pod \"e441e501-09ee-4e3d-a763-c853de09a02c\" (UID: \"e441e501-09ee-4e3d-a763-c853de09a02c\") " Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809616 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "e441e501-09ee-4e3d-a763-c853de09a02c" (UID: "e441e501-09ee-4e3d-a763-c853de09a02c"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809658 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "e441e501-09ee-4e3d-a763-c853de09a02c" (UID: "e441e501-09ee-4e3d-a763-c853de09a02c"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809684 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "e441e501-09ee-4e3d-a763-c853de09a02c" (UID: "e441e501-09ee-4e3d-a763-c853de09a02c"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809715 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-node-log" (OuterVolumeSpecName: "node-log") pod "e441e501-09ee-4e3d-a763-c853de09a02c" (UID: "e441e501-09ee-4e3d-a763-c853de09a02c"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809722 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-kubelet\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809771 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-run-openvswitch\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809800 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-cni-netd\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809826 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/61216c1c-c428-4e6c-83db-19e49744af19-env-overrides\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809901 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-cni-bin\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809936 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-log-socket\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.809997 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.810024 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-etc-openvswitch\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.810092 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-var-lib-openvswitch\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.810148 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/61216c1c-c428-4e6c-83db-19e49744af19-ovn-node-metrics-cert\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.810182 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-slash\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.810235 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8snbb\" (UniqueName: \"kubernetes.io/projected/61216c1c-c428-4e6c-83db-19e49744af19-kube-api-access-8snbb\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.810281 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-run-netns\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.810537 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "e441e501-09ee-4e3d-a763-c853de09a02c" (UID: "e441e501-09ee-4e3d-a763-c853de09a02c"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.810598 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-systemd-units\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.810642 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "e441e501-09ee-4e3d-a763-c853de09a02c" (UID: "e441e501-09ee-4e3d-a763-c853de09a02c"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.810699 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-run-ovn\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.810745 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/61216c1c-c428-4e6c-83db-19e49744af19-ovnkube-script-lib\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.810702 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "e441e501-09ee-4e3d-a763-c853de09a02c" (UID: "e441e501-09ee-4e3d-a763-c853de09a02c"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.810800 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-node-log\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.810827 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-run-ovn-kubernetes\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.810918 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-run-systemd\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.810916 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e441e501-09ee-4e3d-a763-c853de09a02c-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "e441e501-09ee-4e3d-a763-c853de09a02c" (UID: "e441e501-09ee-4e3d-a763-c853de09a02c"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.810960 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e441e501-09ee-4e3d-a763-c853de09a02c-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "e441e501-09ee-4e3d-a763-c853de09a02c" (UID: "e441e501-09ee-4e3d-a763-c853de09a02c"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.811011 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e441e501-09ee-4e3d-a763-c853de09a02c-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "e441e501-09ee-4e3d-a763-c853de09a02c" (UID: "e441e501-09ee-4e3d-a763-c853de09a02c"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.811146 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/61216c1c-c428-4e6c-83db-19e49744af19-ovnkube-config\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.811318 4808 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.811376 4808 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e441e501-09ee-4e3d-a763-c853de09a02c-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.811404 4808 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-node-log\") on node \"crc\" DevicePath \"\"" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.811420 4808 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.811436 4808 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e441e501-09ee-4e3d-a763-c853de09a02c-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.811450 4808 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e441e501-09ee-4e3d-a763-c853de09a02c-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.811465 4808 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.811476 4808 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.811494 4808 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.811523 4808 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.811538 4808 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-log-socket\") on node \"crc\" DevicePath \"\"" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.811552 4808 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.811568 4808 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.811582 4808 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.811597 4808 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.808860 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "e441e501-09ee-4e3d-a763-c853de09a02c" (UID: "e441e501-09ee-4e3d-a763-c853de09a02c"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.816729 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e441e501-09ee-4e3d-a763-c853de09a02c-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "e441e501-09ee-4e3d-a763-c853de09a02c" (UID: "e441e501-09ee-4e3d-a763-c853de09a02c"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.816928 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e441e501-09ee-4e3d-a763-c853de09a02c-kube-api-access-bwkmd" (OuterVolumeSpecName: "kube-api-access-bwkmd") pod "e441e501-09ee-4e3d-a763-c853de09a02c" (UID: "e441e501-09ee-4e3d-a763-c853de09a02c"). InnerVolumeSpecName "kube-api-access-bwkmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.823371 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "e441e501-09ee-4e3d-a763-c853de09a02c" (UID: "e441e501-09ee-4e3d-a763-c853de09a02c"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.911897 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-var-lib-openvswitch\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.911953 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/61216c1c-c428-4e6c-83db-19e49744af19-ovn-node-metrics-cert\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.911976 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-slash\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912004 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8snbb\" (UniqueName: \"kubernetes.io/projected/61216c1c-c428-4e6c-83db-19e49744af19-kube-api-access-8snbb\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912024 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-run-netns\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912046 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-systemd-units\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912067 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-run-ovn\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912086 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/61216c1c-c428-4e6c-83db-19e49744af19-ovnkube-script-lib\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912104 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-run-ovn-kubernetes\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912122 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-run-systemd\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912142 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-node-log\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912175 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/61216c1c-c428-4e6c-83db-19e49744af19-ovnkube-config\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912202 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-kubelet\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912223 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-run-openvswitch\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912263 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-cni-netd\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912281 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/61216c1c-c428-4e6c-83db-19e49744af19-env-overrides\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912310 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-cni-bin\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912333 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-log-socket\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912362 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912385 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-etc-openvswitch\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912434 4808 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912448 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwkmd\" (UniqueName: \"kubernetes.io/projected/e441e501-09ee-4e3d-a763-c853de09a02c-kube-api-access-bwkmd\") on node \"crc\" DevicePath \"\"" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912463 4808 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912475 4808 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e441e501-09ee-4e3d-a763-c853de09a02c-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912489 4808 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e441e501-09ee-4e3d-a763-c853de09a02c-host-slash\") on node \"crc\" DevicePath \"\"" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912542 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-etc-openvswitch\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912534 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-run-systemd\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912586 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-node-log\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912585 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-run-netns\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912656 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-systemd-units\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912673 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-var-lib-openvswitch\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.912715 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-run-ovn\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.913194 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-slash\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.913442 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/61216c1c-c428-4e6c-83db-19e49744af19-ovnkube-config\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.913494 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-kubelet\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.913557 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-run-ovn-kubernetes\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.913797 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/61216c1c-c428-4e6c-83db-19e49744af19-ovnkube-script-lib\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.913823 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-log-socket\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.913798 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-cni-bin\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.913861 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.913903 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-host-cni-netd\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.913863 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61216c1c-c428-4e6c-83db-19e49744af19-run-openvswitch\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.914301 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/61216c1c-c428-4e6c-83db-19e49744af19-env-overrides\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.918923 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/61216c1c-c428-4e6c-83db-19e49744af19-ovn-node-metrics-cert\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.939857 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8snbb\" (UniqueName: \"kubernetes.io/projected/61216c1c-c428-4e6c-83db-19e49744af19-kube-api-access-8snbb\") pod \"ovnkube-node-q6km5\" (UID: \"61216c1c-c428-4e6c-83db-19e49744af19\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:23 crc kubenswrapper[4808]: I1002 16:50:23.997311 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zm46w_e441e501-09ee-4e3d-a763-c853de09a02c/ovnkube-controller/3.log" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.000860 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zm46w_e441e501-09ee-4e3d-a763-c853de09a02c/ovn-acl-logging/0.log" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.001886 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zm46w_e441e501-09ee-4e3d-a763-c853de09a02c/ovn-controller/0.log" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.002677 4808 generic.go:334] "Generic (PLEG): container finished" podID="e441e501-09ee-4e3d-a763-c853de09a02c" containerID="8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419" exitCode=0 Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.002721 4808 generic.go:334] "Generic (PLEG): container finished" podID="e441e501-09ee-4e3d-a763-c853de09a02c" containerID="e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464" exitCode=0 Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.002740 4808 generic.go:334] "Generic (PLEG): container finished" podID="e441e501-09ee-4e3d-a763-c853de09a02c" containerID="15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0" exitCode=0 Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.002758 4808 generic.go:334] "Generic (PLEG): container finished" podID="e441e501-09ee-4e3d-a763-c853de09a02c" containerID="20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6" exitCode=0 Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.002772 4808 generic.go:334] "Generic (PLEG): container finished" podID="e441e501-09ee-4e3d-a763-c853de09a02c" containerID="8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5" exitCode=0 Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.002787 4808 generic.go:334] "Generic (PLEG): container finished" podID="e441e501-09ee-4e3d-a763-c853de09a02c" containerID="964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa" exitCode=0 Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.002792 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.002787 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerDied","Data":"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.002859 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerDied","Data":"e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.002885 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerDied","Data":"15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.002907 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerDied","Data":"20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.002930 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerDied","Data":"8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.002951 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerDied","Data":"964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.002970 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.002991 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003004 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003018 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003032 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003045 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003060 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.002803 4808 generic.go:334] "Generic (PLEG): container finished" podID="e441e501-09ee-4e3d-a763-c853de09a02c" containerID="827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200" exitCode=143 Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003098 4808 generic.go:334] "Generic (PLEG): container finished" podID="e441e501-09ee-4e3d-a763-c853de09a02c" containerID="2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f" exitCode=143 Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003072 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003002 4808 scope.go:117] "RemoveContainer" containerID="8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003198 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003264 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerDied","Data":"827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003301 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003311 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003318 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003330 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003336 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003351 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003357 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003364 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003370 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003376 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003384 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerDied","Data":"2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003395 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003401 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003408 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003418 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003425 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003431 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003437 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003442 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003447 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003453 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003460 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zm46w" event={"ID":"e441e501-09ee-4e3d-a763-c853de09a02c","Type":"ContainerDied","Data":"810f98ce29b314aea121046f13d41ba78b4a3772059264fa2f87dad04eb64289"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003469 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003477 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003483 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003490 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003496 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003502 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003507 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003514 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003520 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.003526 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.006651 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7d9g4_8871f396-b980-46a8-9477-b3be4889638f/kube-multus/2.log" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.007677 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7d9g4_8871f396-b980-46a8-9477-b3be4889638f/kube-multus/1.log" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.007753 4808 generic.go:334] "Generic (PLEG): container finished" podID="8871f396-b980-46a8-9477-b3be4889638f" containerID="6e7bed9dd604e3e5fa64c1b2ee8dba75032ba4b53407915796d9780631b4ec78" exitCode=2 Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.007798 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7d9g4" event={"ID":"8871f396-b980-46a8-9477-b3be4889638f","Type":"ContainerDied","Data":"6e7bed9dd604e3e5fa64c1b2ee8dba75032ba4b53407915796d9780631b4ec78"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.007837 4808 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19f7576c6dc4ec19e5895eb22a1790de5928dbd0c4779a000331f3f45df532c1"} Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.008666 4808 scope.go:117] "RemoveContainer" containerID="6e7bed9dd604e3e5fa64c1b2ee8dba75032ba4b53407915796d9780631b4ec78" Oct 02 16:50:24 crc kubenswrapper[4808]: E1002 16:50:24.009039 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-7d9g4_openshift-multus(8871f396-b980-46a8-9477-b3be4889638f)\"" pod="openshift-multus/multus-7d9g4" podUID="8871f396-b980-46a8-9477-b3be4889638f" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.031563 4808 scope.go:117] "RemoveContainer" containerID="f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.077512 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zm46w"] Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.078578 4808 scope.go:117] "RemoveContainer" containerID="e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.081767 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zm46w"] Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.105410 4808 scope.go:117] "RemoveContainer" containerID="15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.114221 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.124446 4808 scope.go:117] "RemoveContainer" containerID="20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.152053 4808 scope.go:117] "RemoveContainer" containerID="8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.176183 4808 scope.go:117] "RemoveContainer" containerID="964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.199924 4808 scope.go:117] "RemoveContainer" containerID="827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.223382 4808 scope.go:117] "RemoveContainer" containerID="2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.250422 4808 scope.go:117] "RemoveContainer" containerID="99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.282810 4808 scope.go:117] "RemoveContainer" containerID="8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419" Oct 02 16:50:24 crc kubenswrapper[4808]: E1002 16:50:24.283432 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419\": container with ID starting with 8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419 not found: ID does not exist" containerID="8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.283478 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419"} err="failed to get container status \"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419\": rpc error: code = NotFound desc = could not find container \"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419\": container with ID starting with 8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.283510 4808 scope.go:117] "RemoveContainer" containerID="f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97" Oct 02 16:50:24 crc kubenswrapper[4808]: E1002 16:50:24.283964 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97\": container with ID starting with f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97 not found: ID does not exist" containerID="f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.284017 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97"} err="failed to get container status \"f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97\": rpc error: code = NotFound desc = could not find container \"f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97\": container with ID starting with f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.284052 4808 scope.go:117] "RemoveContainer" containerID="e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464" Oct 02 16:50:24 crc kubenswrapper[4808]: E1002 16:50:24.284529 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\": container with ID starting with e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464 not found: ID does not exist" containerID="e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.284585 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464"} err="failed to get container status \"e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\": rpc error: code = NotFound desc = could not find container \"e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\": container with ID starting with e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.284621 4808 scope.go:117] "RemoveContainer" containerID="15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0" Oct 02 16:50:24 crc kubenswrapper[4808]: E1002 16:50:24.284960 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\": container with ID starting with 15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0 not found: ID does not exist" containerID="15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.284992 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0"} err="failed to get container status \"15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\": rpc error: code = NotFound desc = could not find container \"15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\": container with ID starting with 15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.285033 4808 scope.go:117] "RemoveContainer" containerID="20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6" Oct 02 16:50:24 crc kubenswrapper[4808]: E1002 16:50:24.285485 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\": container with ID starting with 20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6 not found: ID does not exist" containerID="20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.285514 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6"} err="failed to get container status \"20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\": rpc error: code = NotFound desc = could not find container \"20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\": container with ID starting with 20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.285533 4808 scope.go:117] "RemoveContainer" containerID="8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5" Oct 02 16:50:24 crc kubenswrapper[4808]: E1002 16:50:24.285891 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\": container with ID starting with 8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5 not found: ID does not exist" containerID="8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.285926 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5"} err="failed to get container status \"8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\": rpc error: code = NotFound desc = could not find container \"8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\": container with ID starting with 8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.285949 4808 scope.go:117] "RemoveContainer" containerID="964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa" Oct 02 16:50:24 crc kubenswrapper[4808]: E1002 16:50:24.286586 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\": container with ID starting with 964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa not found: ID does not exist" containerID="964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.286614 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa"} err="failed to get container status \"964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\": rpc error: code = NotFound desc = could not find container \"964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\": container with ID starting with 964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.286633 4808 scope.go:117] "RemoveContainer" containerID="827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200" Oct 02 16:50:24 crc kubenswrapper[4808]: E1002 16:50:24.287552 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\": container with ID starting with 827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200 not found: ID does not exist" containerID="827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.287577 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200"} err="failed to get container status \"827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\": rpc error: code = NotFound desc = could not find container \"827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\": container with ID starting with 827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.287593 4808 scope.go:117] "RemoveContainer" containerID="2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f" Oct 02 16:50:24 crc kubenswrapper[4808]: E1002 16:50:24.290018 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\": container with ID starting with 2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f not found: ID does not exist" containerID="2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.290055 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f"} err="failed to get container status \"2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\": rpc error: code = NotFound desc = could not find container \"2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\": container with ID starting with 2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.290076 4808 scope.go:117] "RemoveContainer" containerID="99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b" Oct 02 16:50:24 crc kubenswrapper[4808]: E1002 16:50:24.290735 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\": container with ID starting with 99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b not found: ID does not exist" containerID="99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.290761 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b"} err="failed to get container status \"99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\": rpc error: code = NotFound desc = could not find container \"99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\": container with ID starting with 99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.290780 4808 scope.go:117] "RemoveContainer" containerID="8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.291466 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419"} err="failed to get container status \"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419\": rpc error: code = NotFound desc = could not find container \"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419\": container with ID starting with 8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.291492 4808 scope.go:117] "RemoveContainer" containerID="f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.291852 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97"} err="failed to get container status \"f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97\": rpc error: code = NotFound desc = could not find container \"f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97\": container with ID starting with f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.291891 4808 scope.go:117] "RemoveContainer" containerID="e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.292332 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464"} err="failed to get container status \"e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\": rpc error: code = NotFound desc = could not find container \"e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\": container with ID starting with e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.292358 4808 scope.go:117] "RemoveContainer" containerID="15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.292702 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0"} err="failed to get container status \"15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\": rpc error: code = NotFound desc = could not find container \"15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\": container with ID starting with 15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.292726 4808 scope.go:117] "RemoveContainer" containerID="20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.293435 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6"} err="failed to get container status \"20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\": rpc error: code = NotFound desc = could not find container \"20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\": container with ID starting with 20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.293467 4808 scope.go:117] "RemoveContainer" containerID="8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.293768 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5"} err="failed to get container status \"8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\": rpc error: code = NotFound desc = could not find container \"8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\": container with ID starting with 8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.293790 4808 scope.go:117] "RemoveContainer" containerID="964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.294135 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa"} err="failed to get container status \"964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\": rpc error: code = NotFound desc = could not find container \"964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\": container with ID starting with 964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.294158 4808 scope.go:117] "RemoveContainer" containerID="827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.294750 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200"} err="failed to get container status \"827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\": rpc error: code = NotFound desc = could not find container \"827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\": container with ID starting with 827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.294772 4808 scope.go:117] "RemoveContainer" containerID="2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.295377 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f"} err="failed to get container status \"2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\": rpc error: code = NotFound desc = could not find container \"2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\": container with ID starting with 2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.295404 4808 scope.go:117] "RemoveContainer" containerID="99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.298355 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b"} err="failed to get container status \"99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\": rpc error: code = NotFound desc = could not find container \"99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\": container with ID starting with 99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.298380 4808 scope.go:117] "RemoveContainer" containerID="8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.299331 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419"} err="failed to get container status \"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419\": rpc error: code = NotFound desc = could not find container \"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419\": container with ID starting with 8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.299417 4808 scope.go:117] "RemoveContainer" containerID="f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.300364 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97"} err="failed to get container status \"f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97\": rpc error: code = NotFound desc = could not find container \"f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97\": container with ID starting with f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.300422 4808 scope.go:117] "RemoveContainer" containerID="e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.301331 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464"} err="failed to get container status \"e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\": rpc error: code = NotFound desc = could not find container \"e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\": container with ID starting with e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.301385 4808 scope.go:117] "RemoveContainer" containerID="15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.302927 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0"} err="failed to get container status \"15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\": rpc error: code = NotFound desc = could not find container \"15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\": container with ID starting with 15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.302988 4808 scope.go:117] "RemoveContainer" containerID="20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.304651 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6"} err="failed to get container status \"20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\": rpc error: code = NotFound desc = could not find container \"20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\": container with ID starting with 20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.304678 4808 scope.go:117] "RemoveContainer" containerID="8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.305187 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5"} err="failed to get container status \"8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\": rpc error: code = NotFound desc = could not find container \"8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\": container with ID starting with 8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.305211 4808 scope.go:117] "RemoveContainer" containerID="964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.306528 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa"} err="failed to get container status \"964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\": rpc error: code = NotFound desc = could not find container \"964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\": container with ID starting with 964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.306609 4808 scope.go:117] "RemoveContainer" containerID="827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.307120 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200"} err="failed to get container status \"827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\": rpc error: code = NotFound desc = could not find container \"827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\": container with ID starting with 827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.307164 4808 scope.go:117] "RemoveContainer" containerID="2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.307637 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f"} err="failed to get container status \"2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\": rpc error: code = NotFound desc = could not find container \"2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\": container with ID starting with 2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.307681 4808 scope.go:117] "RemoveContainer" containerID="99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.308097 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b"} err="failed to get container status \"99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\": rpc error: code = NotFound desc = could not find container \"99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\": container with ID starting with 99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.308122 4808 scope.go:117] "RemoveContainer" containerID="8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.308549 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419"} err="failed to get container status \"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419\": rpc error: code = NotFound desc = could not find container \"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419\": container with ID starting with 8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.308594 4808 scope.go:117] "RemoveContainer" containerID="f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.309046 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97"} err="failed to get container status \"f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97\": rpc error: code = NotFound desc = could not find container \"f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97\": container with ID starting with f1ccc03cc7dfba650e7bbedb8a42d4960d32351fe52f34f7cb8dba6aecedbd97 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.309073 4808 scope.go:117] "RemoveContainer" containerID="e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.309468 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464"} err="failed to get container status \"e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\": rpc error: code = NotFound desc = could not find container \"e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464\": container with ID starting with e675a3f7f94aa460a90c780efb4ad9482bde334fbaffd310ae30ea06be0ff464 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.309526 4808 scope.go:117] "RemoveContainer" containerID="15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.309937 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0"} err="failed to get container status \"15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\": rpc error: code = NotFound desc = could not find container \"15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0\": container with ID starting with 15d6ae595b9a203b69bc469b3096540a5eae89092eb1046086f15e4da3dfa8e0 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.309963 4808 scope.go:117] "RemoveContainer" containerID="20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.310318 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6"} err="failed to get container status \"20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\": rpc error: code = NotFound desc = could not find container \"20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6\": container with ID starting with 20b749ca5b92565d3039f622eeff74441d6febc727eb2f38961652f5dcd223d6 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.310372 4808 scope.go:117] "RemoveContainer" containerID="8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.310763 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5"} err="failed to get container status \"8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\": rpc error: code = NotFound desc = could not find container \"8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5\": container with ID starting with 8054942c39d1badef6dea508875cc0cb643067a68fed4d0ef791702392eb97d5 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.310786 4808 scope.go:117] "RemoveContainer" containerID="964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.311147 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa"} err="failed to get container status \"964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\": rpc error: code = NotFound desc = could not find container \"964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa\": container with ID starting with 964cd40fd23cdec3d7f5ea64e683c4561f06aa4bf79c8aab19a0edc73a8c54aa not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.311198 4808 scope.go:117] "RemoveContainer" containerID="827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.311632 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200"} err="failed to get container status \"827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\": rpc error: code = NotFound desc = could not find container \"827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200\": container with ID starting with 827f25ead342e209b8232027e2beefd0a228cf52cf0b9d7a9b060498fd259200 not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.311685 4808 scope.go:117] "RemoveContainer" containerID="2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.312145 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f"} err="failed to get container status \"2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\": rpc error: code = NotFound desc = could not find container \"2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f\": container with ID starting with 2b72bbaa860b72c52f00934a112132ffc822f2f6c603110ee5ba2c00c3b6a99f not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.312172 4808 scope.go:117] "RemoveContainer" containerID="99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.312645 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b"} err="failed to get container status \"99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\": rpc error: code = NotFound desc = could not find container \"99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b\": container with ID starting with 99d4d3719de34e665410818c0cce857480a3d8d32ca77abb6d94af459fd3373b not found: ID does not exist" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.312696 4808 scope.go:117] "RemoveContainer" containerID="8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419" Oct 02 16:50:24 crc kubenswrapper[4808]: I1002 16:50:24.313324 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419"} err="failed to get container status \"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419\": rpc error: code = NotFound desc = could not find container \"8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419\": container with ID starting with 8888cfc84100d01e613f45f2ee992b191ae8ad2eefec90fab15aa393c3c50419 not found: ID does not exist" Oct 02 16:50:25 crc kubenswrapper[4808]: I1002 16:50:25.017969 4808 generic.go:334] "Generic (PLEG): container finished" podID="61216c1c-c428-4e6c-83db-19e49744af19" containerID="b0d223607e395c441f86133008f83784102607e97d98617e617a6914cf69fcf7" exitCode=0 Oct 02 16:50:25 crc kubenswrapper[4808]: I1002 16:50:25.018058 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" event={"ID":"61216c1c-c428-4e6c-83db-19e49744af19","Type":"ContainerDied","Data":"b0d223607e395c441f86133008f83784102607e97d98617e617a6914cf69fcf7"} Oct 02 16:50:25 crc kubenswrapper[4808]: I1002 16:50:25.018117 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" event={"ID":"61216c1c-c428-4e6c-83db-19e49744af19","Type":"ContainerStarted","Data":"ebd49862cf8c3b6235ef881f86c2de788fccf2b8ef8e42788846214729d75cce"} Oct 02 16:50:25 crc kubenswrapper[4808]: I1002 16:50:25.406576 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e441e501-09ee-4e3d-a763-c853de09a02c" path="/var/lib/kubelet/pods/e441e501-09ee-4e3d-a763-c853de09a02c/volumes" Oct 02 16:50:26 crc kubenswrapper[4808]: I1002 16:50:26.031140 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" event={"ID":"61216c1c-c428-4e6c-83db-19e49744af19","Type":"ContainerStarted","Data":"2c29eb77d2d8deec933c492b75133798398da051a9ce7ebc422cbb8cd1d8629d"} Oct 02 16:50:26 crc kubenswrapper[4808]: I1002 16:50:26.031763 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" event={"ID":"61216c1c-c428-4e6c-83db-19e49744af19","Type":"ContainerStarted","Data":"d43a0ed23672c5afad5385357d4bb5b9534bc75b83e0df1be8d1280e64a6ee60"} Oct 02 16:50:26 crc kubenswrapper[4808]: I1002 16:50:26.031817 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" event={"ID":"61216c1c-c428-4e6c-83db-19e49744af19","Type":"ContainerStarted","Data":"e19e39586d0fcde4b0c707f542fa76d39a393becf1efa19ca9b703c23827a8f0"} Oct 02 16:50:26 crc kubenswrapper[4808]: I1002 16:50:26.031844 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" event={"ID":"61216c1c-c428-4e6c-83db-19e49744af19","Type":"ContainerStarted","Data":"758c770e6ea972e42793912d58bee7e3374303d2a5d1a7fa8148b974965c614e"} Oct 02 16:50:26 crc kubenswrapper[4808]: I1002 16:50:26.031873 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" event={"ID":"61216c1c-c428-4e6c-83db-19e49744af19","Type":"ContainerStarted","Data":"2a3a48438fe6b6dcb5f5b049fcc189d4a1311e3ccfac0acc9f414700a92140b7"} Oct 02 16:50:27 crc kubenswrapper[4808]: I1002 16:50:27.042762 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" event={"ID":"61216c1c-c428-4e6c-83db-19e49744af19","Type":"ContainerStarted","Data":"debcda895f69ab1c5a30ad6a4d09d0da8d0887e5787e6f6011e2c3aff0f7e0fe"} Oct 02 16:50:29 crc kubenswrapper[4808]: I1002 16:50:29.061139 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" event={"ID":"61216c1c-c428-4e6c-83db-19e49744af19","Type":"ContainerStarted","Data":"1c47ff93914584546b4beb5333dbf17ae78077c239928269140a5f785fa875f9"} Oct 02 16:50:31 crc kubenswrapper[4808]: I1002 16:50:31.079448 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" event={"ID":"61216c1c-c428-4e6c-83db-19e49744af19","Type":"ContainerStarted","Data":"09f2fe943c7f005cf7a55ef0917df66bccd57074279edfb4b438195527d3127b"} Oct 02 16:50:31 crc kubenswrapper[4808]: I1002 16:50:31.079750 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:31 crc kubenswrapper[4808]: I1002 16:50:31.079762 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:31 crc kubenswrapper[4808]: I1002 16:50:31.079772 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:31 crc kubenswrapper[4808]: I1002 16:50:31.119799 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:31 crc kubenswrapper[4808]: I1002 16:50:31.120689 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" podStartSLOduration=8.120672253 podStartE2EDuration="8.120672253s" podCreationTimestamp="2025-10-02 16:50:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:50:31.115614909 +0000 UTC m=+598.441143979" watchObservedRunningTime="2025-10-02 16:50:31.120672253 +0000 UTC m=+598.446201263" Oct 02 16:50:31 crc kubenswrapper[4808]: I1002 16:50:31.128987 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:50:33 crc kubenswrapper[4808]: I1002 16:50:33.592556 4808 scope.go:117] "RemoveContainer" containerID="19f7576c6dc4ec19e5895eb22a1790de5928dbd0c4779a000331f3f45df532c1" Oct 02 16:50:34 crc kubenswrapper[4808]: I1002 16:50:34.102606 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7d9g4_8871f396-b980-46a8-9477-b3be4889638f/kube-multus/2.log" Oct 02 16:50:36 crc kubenswrapper[4808]: I1002 16:50:36.395728 4808 scope.go:117] "RemoveContainer" containerID="6e7bed9dd604e3e5fa64c1b2ee8dba75032ba4b53407915796d9780631b4ec78" Oct 02 16:50:36 crc kubenswrapper[4808]: E1002 16:50:36.396593 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-7d9g4_openshift-multus(8871f396-b980-46a8-9477-b3be4889638f)\"" pod="openshift-multus/multus-7d9g4" podUID="8871f396-b980-46a8-9477-b3be4889638f" Oct 02 16:50:47 crc kubenswrapper[4808]: I1002 16:50:47.396137 4808 scope.go:117] "RemoveContainer" containerID="6e7bed9dd604e3e5fa64c1b2ee8dba75032ba4b53407915796d9780631b4ec78" Oct 02 16:50:48 crc kubenswrapper[4808]: I1002 16:50:48.205499 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7d9g4_8871f396-b980-46a8-9477-b3be4889638f/kube-multus/2.log" Oct 02 16:50:48 crc kubenswrapper[4808]: I1002 16:50:48.205821 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7d9g4" event={"ID":"8871f396-b980-46a8-9477-b3be4889638f","Type":"ContainerStarted","Data":"73910ecd7b7af59e775812a51d1ba79fd3da8d54604a54bd596716c754244ca5"} Oct 02 16:50:48 crc kubenswrapper[4808]: I1002 16:50:48.739667 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 16:50:48 crc kubenswrapper[4808]: I1002 16:50:48.739758 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 16:50:54 crc kubenswrapper[4808]: I1002 16:50:54.156184 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-q6km5" Oct 02 16:51:00 crc kubenswrapper[4808]: I1002 16:51:00.640156 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5"] Oct 02 16:51:00 crc kubenswrapper[4808]: I1002 16:51:00.643039 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5" Oct 02 16:51:00 crc kubenswrapper[4808]: I1002 16:51:00.645593 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 16:51:00 crc kubenswrapper[4808]: I1002 16:51:00.655170 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5"] Oct 02 16:51:00 crc kubenswrapper[4808]: I1002 16:51:00.727989 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htbxq\" (UniqueName: \"kubernetes.io/projected/c19df091-49ac-4606-9cf6-8e9108451371-kube-api-access-htbxq\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5\" (UID: \"c19df091-49ac-4606-9cf6-8e9108451371\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5" Oct 02 16:51:00 crc kubenswrapper[4808]: I1002 16:51:00.728061 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c19df091-49ac-4606-9cf6-8e9108451371-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5\" (UID: \"c19df091-49ac-4606-9cf6-8e9108451371\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5" Oct 02 16:51:00 crc kubenswrapper[4808]: I1002 16:51:00.728092 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c19df091-49ac-4606-9cf6-8e9108451371-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5\" (UID: \"c19df091-49ac-4606-9cf6-8e9108451371\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5" Oct 02 16:51:00 crc kubenswrapper[4808]: I1002 16:51:00.830150 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htbxq\" (UniqueName: \"kubernetes.io/projected/c19df091-49ac-4606-9cf6-8e9108451371-kube-api-access-htbxq\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5\" (UID: \"c19df091-49ac-4606-9cf6-8e9108451371\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5" Oct 02 16:51:00 crc kubenswrapper[4808]: I1002 16:51:00.830227 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c19df091-49ac-4606-9cf6-8e9108451371-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5\" (UID: \"c19df091-49ac-4606-9cf6-8e9108451371\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5" Oct 02 16:51:00 crc kubenswrapper[4808]: I1002 16:51:00.830310 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c19df091-49ac-4606-9cf6-8e9108451371-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5\" (UID: \"c19df091-49ac-4606-9cf6-8e9108451371\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5" Oct 02 16:51:00 crc kubenswrapper[4808]: I1002 16:51:00.831123 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c19df091-49ac-4606-9cf6-8e9108451371-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5\" (UID: \"c19df091-49ac-4606-9cf6-8e9108451371\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5" Oct 02 16:51:00 crc kubenswrapper[4808]: I1002 16:51:00.831423 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c19df091-49ac-4606-9cf6-8e9108451371-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5\" (UID: \"c19df091-49ac-4606-9cf6-8e9108451371\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5" Oct 02 16:51:00 crc kubenswrapper[4808]: I1002 16:51:00.867653 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htbxq\" (UniqueName: \"kubernetes.io/projected/c19df091-49ac-4606-9cf6-8e9108451371-kube-api-access-htbxq\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5\" (UID: \"c19df091-49ac-4606-9cf6-8e9108451371\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5" Oct 02 16:51:00 crc kubenswrapper[4808]: I1002 16:51:00.969948 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5" Oct 02 16:51:01 crc kubenswrapper[4808]: I1002 16:51:01.273050 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5"] Oct 02 16:51:02 crc kubenswrapper[4808]: I1002 16:51:02.291532 4808 generic.go:334] "Generic (PLEG): container finished" podID="c19df091-49ac-4606-9cf6-8e9108451371" containerID="bf1b78a0b1e296b6a5d70c705206ba8de34e2a28917ae20d44b369f745042cdd" exitCode=0 Oct 02 16:51:02 crc kubenswrapper[4808]: I1002 16:51:02.291708 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5" event={"ID":"c19df091-49ac-4606-9cf6-8e9108451371","Type":"ContainerDied","Data":"bf1b78a0b1e296b6a5d70c705206ba8de34e2a28917ae20d44b369f745042cdd"} Oct 02 16:51:02 crc kubenswrapper[4808]: I1002 16:51:02.292042 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5" event={"ID":"c19df091-49ac-4606-9cf6-8e9108451371","Type":"ContainerStarted","Data":"4519dbee9b409c71026e90f212d0a68984bc64e03c162cebf6c0ff5748768f03"} Oct 02 16:51:04 crc kubenswrapper[4808]: I1002 16:51:04.315060 4808 generic.go:334] "Generic (PLEG): container finished" podID="c19df091-49ac-4606-9cf6-8e9108451371" containerID="2aec9548a6839180fd694b3361f85f4ed2ab439b73753db684f2ab8a659c82d8" exitCode=0 Oct 02 16:51:04 crc kubenswrapper[4808]: I1002 16:51:04.315343 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5" event={"ID":"c19df091-49ac-4606-9cf6-8e9108451371","Type":"ContainerDied","Data":"2aec9548a6839180fd694b3361f85f4ed2ab439b73753db684f2ab8a659c82d8"} Oct 02 16:51:05 crc kubenswrapper[4808]: I1002 16:51:05.326645 4808 generic.go:334] "Generic (PLEG): container finished" podID="c19df091-49ac-4606-9cf6-8e9108451371" containerID="5ebe72b1e529d25601e638ae35a4398015964777383653c5bc751ede7638276b" exitCode=0 Oct 02 16:51:05 crc kubenswrapper[4808]: I1002 16:51:05.326703 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5" event={"ID":"c19df091-49ac-4606-9cf6-8e9108451371","Type":"ContainerDied","Data":"5ebe72b1e529d25601e638ae35a4398015964777383653c5bc751ede7638276b"} Oct 02 16:51:06 crc kubenswrapper[4808]: I1002 16:51:06.675357 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5" Oct 02 16:51:06 crc kubenswrapper[4808]: I1002 16:51:06.715124 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htbxq\" (UniqueName: \"kubernetes.io/projected/c19df091-49ac-4606-9cf6-8e9108451371-kube-api-access-htbxq\") pod \"c19df091-49ac-4606-9cf6-8e9108451371\" (UID: \"c19df091-49ac-4606-9cf6-8e9108451371\") " Oct 02 16:51:06 crc kubenswrapper[4808]: I1002 16:51:06.715342 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c19df091-49ac-4606-9cf6-8e9108451371-bundle\") pod \"c19df091-49ac-4606-9cf6-8e9108451371\" (UID: \"c19df091-49ac-4606-9cf6-8e9108451371\") " Oct 02 16:51:06 crc kubenswrapper[4808]: I1002 16:51:06.715390 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c19df091-49ac-4606-9cf6-8e9108451371-util\") pod \"c19df091-49ac-4606-9cf6-8e9108451371\" (UID: \"c19df091-49ac-4606-9cf6-8e9108451371\") " Oct 02 16:51:06 crc kubenswrapper[4808]: I1002 16:51:06.716299 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c19df091-49ac-4606-9cf6-8e9108451371-bundle" (OuterVolumeSpecName: "bundle") pod "c19df091-49ac-4606-9cf6-8e9108451371" (UID: "c19df091-49ac-4606-9cf6-8e9108451371"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:51:06 crc kubenswrapper[4808]: I1002 16:51:06.725364 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c19df091-49ac-4606-9cf6-8e9108451371-kube-api-access-htbxq" (OuterVolumeSpecName: "kube-api-access-htbxq") pod "c19df091-49ac-4606-9cf6-8e9108451371" (UID: "c19df091-49ac-4606-9cf6-8e9108451371"). InnerVolumeSpecName "kube-api-access-htbxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:51:06 crc kubenswrapper[4808]: I1002 16:51:06.817436 4808 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c19df091-49ac-4606-9cf6-8e9108451371-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:51:06 crc kubenswrapper[4808]: I1002 16:51:06.817471 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htbxq\" (UniqueName: \"kubernetes.io/projected/c19df091-49ac-4606-9cf6-8e9108451371-kube-api-access-htbxq\") on node \"crc\" DevicePath \"\"" Oct 02 16:51:06 crc kubenswrapper[4808]: I1002 16:51:06.855136 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c19df091-49ac-4606-9cf6-8e9108451371-util" (OuterVolumeSpecName: "util") pod "c19df091-49ac-4606-9cf6-8e9108451371" (UID: "c19df091-49ac-4606-9cf6-8e9108451371"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:51:06 crc kubenswrapper[4808]: I1002 16:51:06.919115 4808 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c19df091-49ac-4606-9cf6-8e9108451371-util\") on node \"crc\" DevicePath \"\"" Oct 02 16:51:07 crc kubenswrapper[4808]: I1002 16:51:07.343057 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5" event={"ID":"c19df091-49ac-4606-9cf6-8e9108451371","Type":"ContainerDied","Data":"4519dbee9b409c71026e90f212d0a68984bc64e03c162cebf6c0ff5748768f03"} Oct 02 16:51:07 crc kubenswrapper[4808]: I1002 16:51:07.343116 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4519dbee9b409c71026e90f212d0a68984bc64e03c162cebf6c0ff5748768f03" Oct 02 16:51:07 crc kubenswrapper[4808]: I1002 16:51:07.343130 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5" Oct 02 16:51:09 crc kubenswrapper[4808]: I1002 16:51:09.286383 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-998ln"] Oct 02 16:51:09 crc kubenswrapper[4808]: E1002 16:51:09.286823 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19df091-49ac-4606-9cf6-8e9108451371" containerName="extract" Oct 02 16:51:09 crc kubenswrapper[4808]: I1002 16:51:09.286834 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19df091-49ac-4606-9cf6-8e9108451371" containerName="extract" Oct 02 16:51:09 crc kubenswrapper[4808]: E1002 16:51:09.286852 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19df091-49ac-4606-9cf6-8e9108451371" containerName="util" Oct 02 16:51:09 crc kubenswrapper[4808]: I1002 16:51:09.286858 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19df091-49ac-4606-9cf6-8e9108451371" containerName="util" Oct 02 16:51:09 crc kubenswrapper[4808]: E1002 16:51:09.286867 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19df091-49ac-4606-9cf6-8e9108451371" containerName="pull" Oct 02 16:51:09 crc kubenswrapper[4808]: I1002 16:51:09.286874 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19df091-49ac-4606-9cf6-8e9108451371" containerName="pull" Oct 02 16:51:09 crc kubenswrapper[4808]: I1002 16:51:09.286961 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c19df091-49ac-4606-9cf6-8e9108451371" containerName="extract" Oct 02 16:51:09 crc kubenswrapper[4808]: I1002 16:51:09.287534 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-998ln" Oct 02 16:51:09 crc kubenswrapper[4808]: I1002 16:51:09.289175 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 02 16:51:09 crc kubenswrapper[4808]: I1002 16:51:09.289470 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-xfffz" Oct 02 16:51:09 crc kubenswrapper[4808]: I1002 16:51:09.290021 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 02 16:51:09 crc kubenswrapper[4808]: I1002 16:51:09.304201 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-998ln"] Oct 02 16:51:09 crc kubenswrapper[4808]: I1002 16:51:09.353528 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9vvt\" (UniqueName: \"kubernetes.io/projected/b0b5b93b-65b0-4dd5-86a9-2b7151771ff8-kube-api-access-b9vvt\") pod \"nmstate-operator-858ddd8f98-998ln\" (UID: \"b0b5b93b-65b0-4dd5-86a9-2b7151771ff8\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-998ln" Oct 02 16:51:09 crc kubenswrapper[4808]: I1002 16:51:09.455448 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9vvt\" (UniqueName: \"kubernetes.io/projected/b0b5b93b-65b0-4dd5-86a9-2b7151771ff8-kube-api-access-b9vvt\") pod \"nmstate-operator-858ddd8f98-998ln\" (UID: \"b0b5b93b-65b0-4dd5-86a9-2b7151771ff8\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-998ln" Oct 02 16:51:09 crc kubenswrapper[4808]: I1002 16:51:09.490887 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9vvt\" (UniqueName: \"kubernetes.io/projected/b0b5b93b-65b0-4dd5-86a9-2b7151771ff8-kube-api-access-b9vvt\") pod \"nmstate-operator-858ddd8f98-998ln\" (UID: \"b0b5b93b-65b0-4dd5-86a9-2b7151771ff8\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-998ln" Oct 02 16:51:09 crc kubenswrapper[4808]: I1002 16:51:09.606224 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-998ln" Oct 02 16:51:09 crc kubenswrapper[4808]: I1002 16:51:09.830777 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-998ln"] Oct 02 16:51:09 crc kubenswrapper[4808]: W1002 16:51:09.840930 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0b5b93b_65b0_4dd5_86a9_2b7151771ff8.slice/crio-0c462f1cf04c33d888f4c7048c263bb4851199dd3a9cef9678ad706d6a3c4a7f WatchSource:0}: Error finding container 0c462f1cf04c33d888f4c7048c263bb4851199dd3a9cef9678ad706d6a3c4a7f: Status 404 returned error can't find the container with id 0c462f1cf04c33d888f4c7048c263bb4851199dd3a9cef9678ad706d6a3c4a7f Oct 02 16:51:10 crc kubenswrapper[4808]: I1002 16:51:10.361208 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-998ln" event={"ID":"b0b5b93b-65b0-4dd5-86a9-2b7151771ff8","Type":"ContainerStarted","Data":"0c462f1cf04c33d888f4c7048c263bb4851199dd3a9cef9678ad706d6a3c4a7f"} Oct 02 16:51:14 crc kubenswrapper[4808]: I1002 16:51:14.391628 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-998ln" event={"ID":"b0b5b93b-65b0-4dd5-86a9-2b7151771ff8","Type":"ContainerStarted","Data":"0d7c9c9caa1e8cb4711e3ba76100c3cb7dc07745aaf7f609d120220f54e8df3e"} Oct 02 16:51:14 crc kubenswrapper[4808]: I1002 16:51:14.420949 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-998ln" podStartSLOduration=1.098148532 podStartE2EDuration="5.420922112s" podCreationTimestamp="2025-10-02 16:51:09 +0000 UTC" firstStartedPulling="2025-10-02 16:51:09.843756653 +0000 UTC m=+637.169285653" lastFinishedPulling="2025-10-02 16:51:14.166530203 +0000 UTC m=+641.492059233" observedRunningTime="2025-10-02 16:51:14.418109617 +0000 UTC m=+641.743638657" watchObservedRunningTime="2025-10-02 16:51:14.420922112 +0000 UTC m=+641.746451142" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.597066 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-x9xwv"] Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.598747 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-x9xwv" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.603550 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-hfk4t" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.631496 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-tr7dz"] Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.632218 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tr7dz" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.635178 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.651014 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-gljkg"] Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.651750 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-gljkg" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.660595 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-tr7dz"] Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.675737 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-x9xwv"] Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.745613 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqphc\" (UniqueName: \"kubernetes.io/projected/3ee41a88-9b99-4b8c-b985-507fcf86a93e-kube-api-access-kqphc\") pod \"nmstate-metrics-fdff9cb8d-x9xwv\" (UID: \"3ee41a88-9b99-4b8c-b985-507fcf86a93e\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-x9xwv" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.745665 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/3ac3fc4a-ef0b-4066-83e6-ebc7d8364629-dbus-socket\") pod \"nmstate-handler-gljkg\" (UID: \"3ac3fc4a-ef0b-4066-83e6-ebc7d8364629\") " pod="openshift-nmstate/nmstate-handler-gljkg" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.745961 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qj8t8\" (UniqueName: \"kubernetes.io/projected/3ac3fc4a-ef0b-4066-83e6-ebc7d8364629-kube-api-access-qj8t8\") pod \"nmstate-handler-gljkg\" (UID: \"3ac3fc4a-ef0b-4066-83e6-ebc7d8364629\") " pod="openshift-nmstate/nmstate-handler-gljkg" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.747064 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/48ea0327-d95f-42e6-a484-3eb3be93918a-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-tr7dz\" (UID: \"48ea0327-d95f-42e6-a484-3eb3be93918a\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tr7dz" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.747179 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/3ac3fc4a-ef0b-4066-83e6-ebc7d8364629-nmstate-lock\") pod \"nmstate-handler-gljkg\" (UID: \"3ac3fc4a-ef0b-4066-83e6-ebc7d8364629\") " pod="openshift-nmstate/nmstate-handler-gljkg" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.747263 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scgvf\" (UniqueName: \"kubernetes.io/projected/48ea0327-d95f-42e6-a484-3eb3be93918a-kube-api-access-scgvf\") pod \"nmstate-webhook-6cdbc54649-tr7dz\" (UID: \"48ea0327-d95f-42e6-a484-3eb3be93918a\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tr7dz" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.747316 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/3ac3fc4a-ef0b-4066-83e6-ebc7d8364629-ovs-socket\") pod \"nmstate-handler-gljkg\" (UID: \"3ac3fc4a-ef0b-4066-83e6-ebc7d8364629\") " pod="openshift-nmstate/nmstate-handler-gljkg" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.765279 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-96dhm"] Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.766319 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-96dhm" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.768421 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.769203 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.771172 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-jn65m" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.777971 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-96dhm"] Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.849016 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qj8t8\" (UniqueName: \"kubernetes.io/projected/3ac3fc4a-ef0b-4066-83e6-ebc7d8364629-kube-api-access-qj8t8\") pod \"nmstate-handler-gljkg\" (UID: \"3ac3fc4a-ef0b-4066-83e6-ebc7d8364629\") " pod="openshift-nmstate/nmstate-handler-gljkg" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.849090 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5684c3ae-4d4a-4c26-b199-1601088fc825-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-96dhm\" (UID: \"5684c3ae-4d4a-4c26-b199-1601088fc825\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-96dhm" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.849118 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn8m7\" (UniqueName: \"kubernetes.io/projected/5684c3ae-4d4a-4c26-b199-1601088fc825-kube-api-access-sn8m7\") pod \"nmstate-console-plugin-6b874cbd85-96dhm\" (UID: \"5684c3ae-4d4a-4c26-b199-1601088fc825\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-96dhm" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.849138 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5684c3ae-4d4a-4c26-b199-1601088fc825-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-96dhm\" (UID: \"5684c3ae-4d4a-4c26-b199-1601088fc825\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-96dhm" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.849160 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/48ea0327-d95f-42e6-a484-3eb3be93918a-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-tr7dz\" (UID: \"48ea0327-d95f-42e6-a484-3eb3be93918a\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tr7dz" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.849200 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/3ac3fc4a-ef0b-4066-83e6-ebc7d8364629-nmstate-lock\") pod \"nmstate-handler-gljkg\" (UID: \"3ac3fc4a-ef0b-4066-83e6-ebc7d8364629\") " pod="openshift-nmstate/nmstate-handler-gljkg" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.849223 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scgvf\" (UniqueName: \"kubernetes.io/projected/48ea0327-d95f-42e6-a484-3eb3be93918a-kube-api-access-scgvf\") pod \"nmstate-webhook-6cdbc54649-tr7dz\" (UID: \"48ea0327-d95f-42e6-a484-3eb3be93918a\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tr7dz" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.849254 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/3ac3fc4a-ef0b-4066-83e6-ebc7d8364629-ovs-socket\") pod \"nmstate-handler-gljkg\" (UID: \"3ac3fc4a-ef0b-4066-83e6-ebc7d8364629\") " pod="openshift-nmstate/nmstate-handler-gljkg" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.849289 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqphc\" (UniqueName: \"kubernetes.io/projected/3ee41a88-9b99-4b8c-b985-507fcf86a93e-kube-api-access-kqphc\") pod \"nmstate-metrics-fdff9cb8d-x9xwv\" (UID: \"3ee41a88-9b99-4b8c-b985-507fcf86a93e\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-x9xwv" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.849309 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/3ac3fc4a-ef0b-4066-83e6-ebc7d8364629-dbus-socket\") pod \"nmstate-handler-gljkg\" (UID: \"3ac3fc4a-ef0b-4066-83e6-ebc7d8364629\") " pod="openshift-nmstate/nmstate-handler-gljkg" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.849634 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/3ac3fc4a-ef0b-4066-83e6-ebc7d8364629-nmstate-lock\") pod \"nmstate-handler-gljkg\" (UID: \"3ac3fc4a-ef0b-4066-83e6-ebc7d8364629\") " pod="openshift-nmstate/nmstate-handler-gljkg" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.849761 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/3ac3fc4a-ef0b-4066-83e6-ebc7d8364629-dbus-socket\") pod \"nmstate-handler-gljkg\" (UID: \"3ac3fc4a-ef0b-4066-83e6-ebc7d8364629\") " pod="openshift-nmstate/nmstate-handler-gljkg" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.849776 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/3ac3fc4a-ef0b-4066-83e6-ebc7d8364629-ovs-socket\") pod \"nmstate-handler-gljkg\" (UID: \"3ac3fc4a-ef0b-4066-83e6-ebc7d8364629\") " pod="openshift-nmstate/nmstate-handler-gljkg" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.867575 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/48ea0327-d95f-42e6-a484-3eb3be93918a-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-tr7dz\" (UID: \"48ea0327-d95f-42e6-a484-3eb3be93918a\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tr7dz" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.868962 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqphc\" (UniqueName: \"kubernetes.io/projected/3ee41a88-9b99-4b8c-b985-507fcf86a93e-kube-api-access-kqphc\") pod \"nmstate-metrics-fdff9cb8d-x9xwv\" (UID: \"3ee41a88-9b99-4b8c-b985-507fcf86a93e\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-x9xwv" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.874162 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scgvf\" (UniqueName: \"kubernetes.io/projected/48ea0327-d95f-42e6-a484-3eb3be93918a-kube-api-access-scgvf\") pod \"nmstate-webhook-6cdbc54649-tr7dz\" (UID: \"48ea0327-d95f-42e6-a484-3eb3be93918a\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tr7dz" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.882645 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qj8t8\" (UniqueName: \"kubernetes.io/projected/3ac3fc4a-ef0b-4066-83e6-ebc7d8364629-kube-api-access-qj8t8\") pod \"nmstate-handler-gljkg\" (UID: \"3ac3fc4a-ef0b-4066-83e6-ebc7d8364629\") " pod="openshift-nmstate/nmstate-handler-gljkg" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.917770 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-x9xwv" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.953100 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5684c3ae-4d4a-4c26-b199-1601088fc825-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-96dhm\" (UID: \"5684c3ae-4d4a-4c26-b199-1601088fc825\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-96dhm" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.953338 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn8m7\" (UniqueName: \"kubernetes.io/projected/5684c3ae-4d4a-4c26-b199-1601088fc825-kube-api-access-sn8m7\") pod \"nmstate-console-plugin-6b874cbd85-96dhm\" (UID: \"5684c3ae-4d4a-4c26-b199-1601088fc825\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-96dhm" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.953434 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5684c3ae-4d4a-4c26-b199-1601088fc825-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-96dhm\" (UID: \"5684c3ae-4d4a-4c26-b199-1601088fc825\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-96dhm" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.954454 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5684c3ae-4d4a-4c26-b199-1601088fc825-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-96dhm\" (UID: \"5684c3ae-4d4a-4c26-b199-1601088fc825\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-96dhm" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.956887 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tr7dz" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.959829 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7fbcd55fbb-kwk9w"] Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.960522 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.968460 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5684c3ae-4d4a-4c26-b199-1601088fc825-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-96dhm\" (UID: \"5684c3ae-4d4a-4c26-b199-1601088fc825\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-96dhm" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.968741 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-gljkg" Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.974721 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7fbcd55fbb-kwk9w"] Oct 02 16:51:15 crc kubenswrapper[4808]: I1002 16:51:15.976829 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn8m7\" (UniqueName: \"kubernetes.io/projected/5684c3ae-4d4a-4c26-b199-1601088fc825-kube-api-access-sn8m7\") pod \"nmstate-console-plugin-6b874cbd85-96dhm\" (UID: \"5684c3ae-4d4a-4c26-b199-1601088fc825\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-96dhm" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.085588 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-96dhm" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.157712 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8m8p\" (UniqueName: \"kubernetes.io/projected/20975205-942b-4a5f-bc33-a0a70d2b6391-kube-api-access-n8m8p\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.157751 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/20975205-942b-4a5f-bc33-a0a70d2b6391-console-serving-cert\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.157781 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/20975205-942b-4a5f-bc33-a0a70d2b6391-console-oauth-config\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.157806 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/20975205-942b-4a5f-bc33-a0a70d2b6391-trusted-ca-bundle\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.157837 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/20975205-942b-4a5f-bc33-a0a70d2b6391-oauth-serving-cert\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.157853 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/20975205-942b-4a5f-bc33-a0a70d2b6391-service-ca\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.157891 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/20975205-942b-4a5f-bc33-a0a70d2b6391-console-config\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.232615 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-tr7dz"] Oct 02 16:51:16 crc kubenswrapper[4808]: W1002 16:51:16.236788 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48ea0327_d95f_42e6_a484_3eb3be93918a.slice/crio-5f69575eb201dfd6035cfa62ceb152f486d548309fd5d53d7ed86a1e53d79c56 WatchSource:0}: Error finding container 5f69575eb201dfd6035cfa62ceb152f486d548309fd5d53d7ed86a1e53d79c56: Status 404 returned error can't find the container with id 5f69575eb201dfd6035cfa62ceb152f486d548309fd5d53d7ed86a1e53d79c56 Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.258882 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/20975205-942b-4a5f-bc33-a0a70d2b6391-console-config\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.259584 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8m8p\" (UniqueName: \"kubernetes.io/projected/20975205-942b-4a5f-bc33-a0a70d2b6391-kube-api-access-n8m8p\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.259650 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/20975205-942b-4a5f-bc33-a0a70d2b6391-console-serving-cert\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.260699 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/20975205-942b-4a5f-bc33-a0a70d2b6391-console-oauth-config\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.260765 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/20975205-942b-4a5f-bc33-a0a70d2b6391-trusted-ca-bundle\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.260835 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/20975205-942b-4a5f-bc33-a0a70d2b6391-oauth-serving-cert\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.260867 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/20975205-942b-4a5f-bc33-a0a70d2b6391-service-ca\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.261585 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/20975205-942b-4a5f-bc33-a0a70d2b6391-oauth-serving-cert\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.261954 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/20975205-942b-4a5f-bc33-a0a70d2b6391-service-ca\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.262510 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/20975205-942b-4a5f-bc33-a0a70d2b6391-console-config\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.266653 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/20975205-942b-4a5f-bc33-a0a70d2b6391-console-serving-cert\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.267980 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/20975205-942b-4a5f-bc33-a0a70d2b6391-trusted-ca-bundle\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.268695 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/20975205-942b-4a5f-bc33-a0a70d2b6391-console-oauth-config\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.276217 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8m8p\" (UniqueName: \"kubernetes.io/projected/20975205-942b-4a5f-bc33-a0a70d2b6391-kube-api-access-n8m8p\") pod \"console-7fbcd55fbb-kwk9w\" (UID: \"20975205-942b-4a5f-bc33-a0a70d2b6391\") " pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.288103 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-96dhm"] Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.299171 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.406097 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-gljkg" event={"ID":"3ac3fc4a-ef0b-4066-83e6-ebc7d8364629","Type":"ContainerStarted","Data":"9927b4fed0bb6bddd701d66621dd4ad8fce6464590b8b0fccf204455417b8526"} Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.406998 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tr7dz" event={"ID":"48ea0327-d95f-42e6-a484-3eb3be93918a","Type":"ContainerStarted","Data":"5f69575eb201dfd6035cfa62ceb152f486d548309fd5d53d7ed86a1e53d79c56"} Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.407921 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-96dhm" event={"ID":"5684c3ae-4d4a-4c26-b199-1601088fc825","Type":"ContainerStarted","Data":"c3795fb615903e9e400ca48c8273e66d93affa161fc181daf24f9c7d160b75c3"} Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.436859 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-x9xwv"] Oct 02 16:51:16 crc kubenswrapper[4808]: W1002 16:51:16.440883 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ee41a88_9b99_4b8c_b985_507fcf86a93e.slice/crio-b2e167de56c8ed67e05da879fec68d04ae58bbb90f99ea8947edbae517038667 WatchSource:0}: Error finding container b2e167de56c8ed67e05da879fec68d04ae58bbb90f99ea8947edbae517038667: Status 404 returned error can't find the container with id b2e167de56c8ed67e05da879fec68d04ae58bbb90f99ea8947edbae517038667 Oct 02 16:51:16 crc kubenswrapper[4808]: I1002 16:51:16.758260 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7fbcd55fbb-kwk9w"] Oct 02 16:51:16 crc kubenswrapper[4808]: W1002 16:51:16.770988 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20975205_942b_4a5f_bc33_a0a70d2b6391.slice/crio-2f1f33aeb5aebb1d6a82d05af09357cc9b84c446acfbfc55ecf96a18a8f55caf WatchSource:0}: Error finding container 2f1f33aeb5aebb1d6a82d05af09357cc9b84c446acfbfc55ecf96a18a8f55caf: Status 404 returned error can't find the container with id 2f1f33aeb5aebb1d6a82d05af09357cc9b84c446acfbfc55ecf96a18a8f55caf Oct 02 16:51:17 crc kubenswrapper[4808]: I1002 16:51:17.415960 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7fbcd55fbb-kwk9w" event={"ID":"20975205-942b-4a5f-bc33-a0a70d2b6391","Type":"ContainerStarted","Data":"ec92c229a58680e1d965920014d7959b1c01470bfe19571ad4481039c2ec5788"} Oct 02 16:51:17 crc kubenswrapper[4808]: I1002 16:51:17.416376 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7fbcd55fbb-kwk9w" event={"ID":"20975205-942b-4a5f-bc33-a0a70d2b6391","Type":"ContainerStarted","Data":"2f1f33aeb5aebb1d6a82d05af09357cc9b84c446acfbfc55ecf96a18a8f55caf"} Oct 02 16:51:17 crc kubenswrapper[4808]: I1002 16:51:17.417552 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-x9xwv" event={"ID":"3ee41a88-9b99-4b8c-b985-507fcf86a93e","Type":"ContainerStarted","Data":"b2e167de56c8ed67e05da879fec68d04ae58bbb90f99ea8947edbae517038667"} Oct 02 16:51:17 crc kubenswrapper[4808]: I1002 16:51:17.442122 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7fbcd55fbb-kwk9w" podStartSLOduration=2.442101052 podStartE2EDuration="2.442101052s" podCreationTimestamp="2025-10-02 16:51:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:51:17.433965296 +0000 UTC m=+644.759494296" watchObservedRunningTime="2025-10-02 16:51:17.442101052 +0000 UTC m=+644.767630052" Oct 02 16:51:18 crc kubenswrapper[4808]: I1002 16:51:18.739960 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 16:51:18 crc kubenswrapper[4808]: I1002 16:51:18.740330 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 16:51:18 crc kubenswrapper[4808]: I1002 16:51:18.740387 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 16:51:18 crc kubenswrapper[4808]: I1002 16:51:18.741086 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a9406662e09c7be4b59f910ebd57fef50f5b64e0992e5ddee0ca92a0e3d8521a"} pod="openshift-machine-config-operator/machine-config-daemon-7z66r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 16:51:18 crc kubenswrapper[4808]: I1002 16:51:18.741139 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" containerID="cri-o://a9406662e09c7be4b59f910ebd57fef50f5b64e0992e5ddee0ca92a0e3d8521a" gracePeriod=600 Oct 02 16:51:19 crc kubenswrapper[4808]: I1002 16:51:19.446359 4808 generic.go:334] "Generic (PLEG): container finished" podID="f83ce425-101d-4489-94a4-5c256eb29328" containerID="a9406662e09c7be4b59f910ebd57fef50f5b64e0992e5ddee0ca92a0e3d8521a" exitCode=0 Oct 02 16:51:19 crc kubenswrapper[4808]: I1002 16:51:19.446443 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" event={"ID":"f83ce425-101d-4489-94a4-5c256eb29328","Type":"ContainerDied","Data":"a9406662e09c7be4b59f910ebd57fef50f5b64e0992e5ddee0ca92a0e3d8521a"} Oct 02 16:51:19 crc kubenswrapper[4808]: I1002 16:51:19.447280 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" event={"ID":"f83ce425-101d-4489-94a4-5c256eb29328","Type":"ContainerStarted","Data":"8d253701a5ddc087ee8a36b225b0284b63257d65500bbf63a65e988cba6e1fa1"} Oct 02 16:51:19 crc kubenswrapper[4808]: I1002 16:51:19.447305 4808 scope.go:117] "RemoveContainer" containerID="f801630a113a03b3e035f8940f01d306b49a7227476e62d1552f3229787c1924" Oct 02 16:51:19 crc kubenswrapper[4808]: I1002 16:51:19.449048 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-x9xwv" event={"ID":"3ee41a88-9b99-4b8c-b985-507fcf86a93e","Type":"ContainerStarted","Data":"add597113f7c987ce5deab6152df85d65480e42d380a06f319af261ab403255d"} Oct 02 16:51:19 crc kubenswrapper[4808]: I1002 16:51:19.452091 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-gljkg" event={"ID":"3ac3fc4a-ef0b-4066-83e6-ebc7d8364629","Type":"ContainerStarted","Data":"1d010e837f844de764f0ecff6b02e7b696850e52cd8ccb0fb1bf0a4b0bb7fb1e"} Oct 02 16:51:19 crc kubenswrapper[4808]: I1002 16:51:19.452141 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-gljkg" Oct 02 16:51:19 crc kubenswrapper[4808]: I1002 16:51:19.453275 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tr7dz" event={"ID":"48ea0327-d95f-42e6-a484-3eb3be93918a","Type":"ContainerStarted","Data":"763976e1cfdc7beccbaafd0315e0515ae90c5e58e7d93514205c5ed6c8d4ef48"} Oct 02 16:51:19 crc kubenswrapper[4808]: I1002 16:51:19.453479 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tr7dz" Oct 02 16:51:19 crc kubenswrapper[4808]: I1002 16:51:19.505107 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-gljkg" podStartSLOduration=2.094995317 podStartE2EDuration="4.50508528s" podCreationTimestamp="2025-10-02 16:51:15 +0000 UTC" firstStartedPulling="2025-10-02 16:51:16.031178907 +0000 UTC m=+643.356707907" lastFinishedPulling="2025-10-02 16:51:18.44126886 +0000 UTC m=+645.766797870" observedRunningTime="2025-10-02 16:51:19.482313766 +0000 UTC m=+646.807842766" watchObservedRunningTime="2025-10-02 16:51:19.50508528 +0000 UTC m=+646.830614280" Oct 02 16:51:19 crc kubenswrapper[4808]: I1002 16:51:19.516125 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tr7dz" podStartSLOduration=2.2877883629999998 podStartE2EDuration="4.516098721s" podCreationTimestamp="2025-10-02 16:51:15 +0000 UTC" firstStartedPulling="2025-10-02 16:51:16.239151106 +0000 UTC m=+643.564680106" lastFinishedPulling="2025-10-02 16:51:18.467461424 +0000 UTC m=+645.792990464" observedRunningTime="2025-10-02 16:51:19.504591646 +0000 UTC m=+646.830120666" watchObservedRunningTime="2025-10-02 16:51:19.516098721 +0000 UTC m=+646.841627721" Oct 02 16:51:20 crc kubenswrapper[4808]: I1002 16:51:20.466802 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-96dhm" event={"ID":"5684c3ae-4d4a-4c26-b199-1601088fc825","Type":"ContainerStarted","Data":"d4d4fc454767c60a3c9df17481ddec65d45a06563d6d66d1c5a69fc13132cf42"} Oct 02 16:51:20 crc kubenswrapper[4808]: I1002 16:51:20.485690 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-96dhm" podStartSLOduration=2.29669251 podStartE2EDuration="5.485667315s" podCreationTimestamp="2025-10-02 16:51:15 +0000 UTC" firstStartedPulling="2025-10-02 16:51:16.294655526 +0000 UTC m=+643.620184536" lastFinishedPulling="2025-10-02 16:51:19.483630341 +0000 UTC m=+646.809159341" observedRunningTime="2025-10-02 16:51:20.483804666 +0000 UTC m=+647.809333706" watchObservedRunningTime="2025-10-02 16:51:20.485667315 +0000 UTC m=+647.811196305" Oct 02 16:51:21 crc kubenswrapper[4808]: I1002 16:51:21.479674 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-x9xwv" event={"ID":"3ee41a88-9b99-4b8c-b985-507fcf86a93e","Type":"ContainerStarted","Data":"70b33a5740af59dd334f7eb6d651faba0c7556f89ca35899e65ce19d8dfa4454"} Oct 02 16:51:21 crc kubenswrapper[4808]: I1002 16:51:21.510921 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-x9xwv" podStartSLOduration=2.216939377 podStartE2EDuration="6.510891844s" podCreationTimestamp="2025-10-02 16:51:15 +0000 UTC" firstStartedPulling="2025-10-02 16:51:16.443405136 +0000 UTC m=+643.768934136" lastFinishedPulling="2025-10-02 16:51:20.737357592 +0000 UTC m=+648.062886603" observedRunningTime="2025-10-02 16:51:21.50547882 +0000 UTC m=+648.831007850" watchObservedRunningTime="2025-10-02 16:51:21.510891844 +0000 UTC m=+648.836420874" Oct 02 16:51:25 crc kubenswrapper[4808]: I1002 16:51:25.990273 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-gljkg" Oct 02 16:51:26 crc kubenswrapper[4808]: I1002 16:51:26.300144 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:26 crc kubenswrapper[4808]: I1002 16:51:26.300278 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:26 crc kubenswrapper[4808]: I1002 16:51:26.306034 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:26 crc kubenswrapper[4808]: I1002 16:51:26.524459 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7fbcd55fbb-kwk9w" Oct 02 16:51:26 crc kubenswrapper[4808]: I1002 16:51:26.592053 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-dlddx"] Oct 02 16:51:35 crc kubenswrapper[4808]: I1002 16:51:35.967420 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tr7dz" Oct 02 16:51:51 crc kubenswrapper[4808]: I1002 16:51:51.649685 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-dlddx" podUID="c58d3855-bb32-4bb3-94ae-53c49209655e" containerName="console" containerID="cri-o://747c3304769acdbfac095d6353f25b73ea9de1393a4aea5f785e491d56d1445c" gracePeriod=15 Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.131997 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-dlddx_c58d3855-bb32-4bb3-94ae-53c49209655e/console/0.log" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.132114 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.279280 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-oauth-serving-cert\") pod \"c58d3855-bb32-4bb3-94ae-53c49209655e\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.279330 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c58d3855-bb32-4bb3-94ae-53c49209655e-console-serving-cert\") pod \"c58d3855-bb32-4bb3-94ae-53c49209655e\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.279354 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c58d3855-bb32-4bb3-94ae-53c49209655e-console-oauth-config\") pod \"c58d3855-bb32-4bb3-94ae-53c49209655e\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.279382 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-trusted-ca-bundle\") pod \"c58d3855-bb32-4bb3-94ae-53c49209655e\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.279463 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-service-ca\") pod \"c58d3855-bb32-4bb3-94ae-53c49209655e\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.279488 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-console-config\") pod \"c58d3855-bb32-4bb3-94ae-53c49209655e\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.279532 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lv5c8\" (UniqueName: \"kubernetes.io/projected/c58d3855-bb32-4bb3-94ae-53c49209655e-kube-api-access-lv5c8\") pod \"c58d3855-bb32-4bb3-94ae-53c49209655e\" (UID: \"c58d3855-bb32-4bb3-94ae-53c49209655e\") " Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.281049 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "c58d3855-bb32-4bb3-94ae-53c49209655e" (UID: "c58d3855-bb32-4bb3-94ae-53c49209655e"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.281097 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-service-ca" (OuterVolumeSpecName: "service-ca") pod "c58d3855-bb32-4bb3-94ae-53c49209655e" (UID: "c58d3855-bb32-4bb3-94ae-53c49209655e"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.281118 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-console-config" (OuterVolumeSpecName: "console-config") pod "c58d3855-bb32-4bb3-94ae-53c49209655e" (UID: "c58d3855-bb32-4bb3-94ae-53c49209655e"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.281354 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "c58d3855-bb32-4bb3-94ae-53c49209655e" (UID: "c58d3855-bb32-4bb3-94ae-53c49209655e"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.285413 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c58d3855-bb32-4bb3-94ae-53c49209655e-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "c58d3855-bb32-4bb3-94ae-53c49209655e" (UID: "c58d3855-bb32-4bb3-94ae-53c49209655e"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.286687 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c58d3855-bb32-4bb3-94ae-53c49209655e-kube-api-access-lv5c8" (OuterVolumeSpecName: "kube-api-access-lv5c8") pod "c58d3855-bb32-4bb3-94ae-53c49209655e" (UID: "c58d3855-bb32-4bb3-94ae-53c49209655e"). InnerVolumeSpecName "kube-api-access-lv5c8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.286854 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c58d3855-bb32-4bb3-94ae-53c49209655e-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "c58d3855-bb32-4bb3-94ae-53c49209655e" (UID: "c58d3855-bb32-4bb3-94ae-53c49209655e"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.381082 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lv5c8\" (UniqueName: \"kubernetes.io/projected/c58d3855-bb32-4bb3-94ae-53c49209655e-kube-api-access-lv5c8\") on node \"crc\" DevicePath \"\"" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.381715 4808 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.381735 4808 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c58d3855-bb32-4bb3-94ae-53c49209655e-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.381749 4808 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c58d3855-bb32-4bb3-94ae-53c49209655e-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.381769 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.381786 4808 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.381802 4808 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c58d3855-bb32-4bb3-94ae-53c49209655e-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.716540 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-dlddx_c58d3855-bb32-4bb3-94ae-53c49209655e/console/0.log" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.716595 4808 generic.go:334] "Generic (PLEG): container finished" podID="c58d3855-bb32-4bb3-94ae-53c49209655e" containerID="747c3304769acdbfac095d6353f25b73ea9de1393a4aea5f785e491d56d1445c" exitCode=2 Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.716624 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-dlddx" event={"ID":"c58d3855-bb32-4bb3-94ae-53c49209655e","Type":"ContainerDied","Data":"747c3304769acdbfac095d6353f25b73ea9de1393a4aea5f785e491d56d1445c"} Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.716651 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-dlddx" event={"ID":"c58d3855-bb32-4bb3-94ae-53c49209655e","Type":"ContainerDied","Data":"78ca84097692742cb9b739f329a421350f052830b92da98d0bb31eb70215430f"} Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.716669 4808 scope.go:117] "RemoveContainer" containerID="747c3304769acdbfac095d6353f25b73ea9de1393a4aea5f785e491d56d1445c" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.716737 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-dlddx" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.740020 4808 scope.go:117] "RemoveContainer" containerID="747c3304769acdbfac095d6353f25b73ea9de1393a4aea5f785e491d56d1445c" Oct 02 16:51:52 crc kubenswrapper[4808]: E1002 16:51:52.740700 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"747c3304769acdbfac095d6353f25b73ea9de1393a4aea5f785e491d56d1445c\": container with ID starting with 747c3304769acdbfac095d6353f25b73ea9de1393a4aea5f785e491d56d1445c not found: ID does not exist" containerID="747c3304769acdbfac095d6353f25b73ea9de1393a4aea5f785e491d56d1445c" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.740791 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"747c3304769acdbfac095d6353f25b73ea9de1393a4aea5f785e491d56d1445c"} err="failed to get container status \"747c3304769acdbfac095d6353f25b73ea9de1393a4aea5f785e491d56d1445c\": rpc error: code = NotFound desc = could not find container \"747c3304769acdbfac095d6353f25b73ea9de1393a4aea5f785e491d56d1445c\": container with ID starting with 747c3304769acdbfac095d6353f25b73ea9de1393a4aea5f785e491d56d1445c not found: ID does not exist" Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.762616 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-dlddx"] Oct 02 16:51:52 crc kubenswrapper[4808]: I1002 16:51:52.765728 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-dlddx"] Oct 02 16:51:53 crc kubenswrapper[4808]: I1002 16:51:53.408448 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c58d3855-bb32-4bb3-94ae-53c49209655e" path="/var/lib/kubelet/pods/c58d3855-bb32-4bb3-94ae-53c49209655e/volumes" Oct 02 16:51:53 crc kubenswrapper[4808]: I1002 16:51:53.852823 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl"] Oct 02 16:51:53 crc kubenswrapper[4808]: E1002 16:51:53.853311 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c58d3855-bb32-4bb3-94ae-53c49209655e" containerName="console" Oct 02 16:51:53 crc kubenswrapper[4808]: I1002 16:51:53.853325 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c58d3855-bb32-4bb3-94ae-53c49209655e" containerName="console" Oct 02 16:51:53 crc kubenswrapper[4808]: I1002 16:51:53.853463 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c58d3855-bb32-4bb3-94ae-53c49209655e" containerName="console" Oct 02 16:51:53 crc kubenswrapper[4808]: I1002 16:51:53.854321 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl" Oct 02 16:51:53 crc kubenswrapper[4808]: I1002 16:51:53.856996 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 16:51:53 crc kubenswrapper[4808]: I1002 16:51:53.865944 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl"] Oct 02 16:51:54 crc kubenswrapper[4808]: I1002 16:51:54.008132 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl\" (UID: \"f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl" Oct 02 16:51:54 crc kubenswrapper[4808]: I1002 16:51:54.008218 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctkbm\" (UniqueName: \"kubernetes.io/projected/f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6-kube-api-access-ctkbm\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl\" (UID: \"f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl" Oct 02 16:51:54 crc kubenswrapper[4808]: I1002 16:51:54.008304 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl\" (UID: \"f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl" Oct 02 16:51:54 crc kubenswrapper[4808]: I1002 16:51:54.109668 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl\" (UID: \"f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl" Oct 02 16:51:54 crc kubenswrapper[4808]: I1002 16:51:54.109760 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctkbm\" (UniqueName: \"kubernetes.io/projected/f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6-kube-api-access-ctkbm\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl\" (UID: \"f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl" Oct 02 16:51:54 crc kubenswrapper[4808]: I1002 16:51:54.109818 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl\" (UID: \"f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl" Oct 02 16:51:54 crc kubenswrapper[4808]: I1002 16:51:54.110447 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl\" (UID: \"f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl" Oct 02 16:51:54 crc kubenswrapper[4808]: I1002 16:51:54.110539 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl\" (UID: \"f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl" Oct 02 16:51:54 crc kubenswrapper[4808]: I1002 16:51:54.145030 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctkbm\" (UniqueName: \"kubernetes.io/projected/f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6-kube-api-access-ctkbm\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl\" (UID: \"f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl" Oct 02 16:51:54 crc kubenswrapper[4808]: I1002 16:51:54.192649 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl" Oct 02 16:51:54 crc kubenswrapper[4808]: I1002 16:51:54.665288 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl"] Oct 02 16:51:54 crc kubenswrapper[4808]: I1002 16:51:54.738013 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl" event={"ID":"f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6","Type":"ContainerStarted","Data":"a7ec8053c5804e36b657759aad2242523224e9a50717caf5c7260d9085afe129"} Oct 02 16:51:55 crc kubenswrapper[4808]: I1002 16:51:55.748323 4808 generic.go:334] "Generic (PLEG): container finished" podID="f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6" containerID="e7ae73f7e2ff39b0a87c64d0c45b51fb160196190260cc1e79f6bc5cee0dda9b" exitCode=0 Oct 02 16:51:55 crc kubenswrapper[4808]: I1002 16:51:55.748494 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl" event={"ID":"f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6","Type":"ContainerDied","Data":"e7ae73f7e2ff39b0a87c64d0c45b51fb160196190260cc1e79f6bc5cee0dda9b"} Oct 02 16:51:59 crc kubenswrapper[4808]: I1002 16:51:59.783666 4808 generic.go:334] "Generic (PLEG): container finished" podID="f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6" containerID="52bb981f088c2166e039ed5b611ef404b293b0a173ff26faf971734c358fdb24" exitCode=0 Oct 02 16:51:59 crc kubenswrapper[4808]: I1002 16:51:59.783766 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl" event={"ID":"f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6","Type":"ContainerDied","Data":"52bb981f088c2166e039ed5b611ef404b293b0a173ff26faf971734c358fdb24"} Oct 02 16:52:00 crc kubenswrapper[4808]: I1002 16:52:00.793498 4808 generic.go:334] "Generic (PLEG): container finished" podID="f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6" containerID="639a5827fc34963819753942eaf4bb6442e5b7d729021f36d4fd1e0f8658fe99" exitCode=0 Oct 02 16:52:00 crc kubenswrapper[4808]: I1002 16:52:00.793599 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl" event={"ID":"f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6","Type":"ContainerDied","Data":"639a5827fc34963819753942eaf4bb6442e5b7d729021f36d4fd1e0f8658fe99"} Oct 02 16:52:02 crc kubenswrapper[4808]: I1002 16:52:02.128895 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl" Oct 02 16:52:02 crc kubenswrapper[4808]: I1002 16:52:02.237317 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6-util\") pod \"f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6\" (UID: \"f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6\") " Oct 02 16:52:02 crc kubenswrapper[4808]: I1002 16:52:02.237521 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6-bundle\") pod \"f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6\" (UID: \"f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6\") " Oct 02 16:52:02 crc kubenswrapper[4808]: I1002 16:52:02.237615 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctkbm\" (UniqueName: \"kubernetes.io/projected/f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6-kube-api-access-ctkbm\") pod \"f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6\" (UID: \"f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6\") " Oct 02 16:52:02 crc kubenswrapper[4808]: I1002 16:52:02.239313 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6-bundle" (OuterVolumeSpecName: "bundle") pod "f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6" (UID: "f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:52:02 crc kubenswrapper[4808]: I1002 16:52:02.246980 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6-kube-api-access-ctkbm" (OuterVolumeSpecName: "kube-api-access-ctkbm") pod "f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6" (UID: "f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6"). InnerVolumeSpecName "kube-api-access-ctkbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:52:02 crc kubenswrapper[4808]: I1002 16:52:02.261656 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6-util" (OuterVolumeSpecName: "util") pod "f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6" (UID: "f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:52:02 crc kubenswrapper[4808]: I1002 16:52:02.338656 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctkbm\" (UniqueName: \"kubernetes.io/projected/f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6-kube-api-access-ctkbm\") on node \"crc\" DevicePath \"\"" Oct 02 16:52:02 crc kubenswrapper[4808]: I1002 16:52:02.338731 4808 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6-util\") on node \"crc\" DevicePath \"\"" Oct 02 16:52:02 crc kubenswrapper[4808]: I1002 16:52:02.338750 4808 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:52:02 crc kubenswrapper[4808]: I1002 16:52:02.814403 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl" event={"ID":"f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6","Type":"ContainerDied","Data":"a7ec8053c5804e36b657759aad2242523224e9a50717caf5c7260d9085afe129"} Oct 02 16:52:02 crc kubenswrapper[4808]: I1002 16:52:02.814495 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl" Oct 02 16:52:02 crc kubenswrapper[4808]: I1002 16:52:02.814499 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7ec8053c5804e36b657759aad2242523224e9a50717caf5c7260d9085afe129" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.591218 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d89bdf8b7-wfkk6"] Oct 02 16:52:12 crc kubenswrapper[4808]: E1002 16:52:12.592179 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6" containerName="extract" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.592196 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6" containerName="extract" Oct 02 16:52:12 crc kubenswrapper[4808]: E1002 16:52:12.592221 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6" containerName="util" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.592246 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6" containerName="util" Oct 02 16:52:12 crc kubenswrapper[4808]: E1002 16:52:12.592264 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6" containerName="pull" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.592272 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6" containerName="pull" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.592564 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6" containerName="extract" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.593317 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5d89bdf8b7-wfkk6" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.601829 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-rvt6m" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.601977 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.602203 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.602544 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.607767 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d89bdf8b7-wfkk6"] Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.610389 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.678957 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b8be9f85-599f-4e47-99f5-c4e3d9ef27e9-webhook-cert\") pod \"metallb-operator-controller-manager-5d89bdf8b7-wfkk6\" (UID: \"b8be9f85-599f-4e47-99f5-c4e3d9ef27e9\") " pod="metallb-system/metallb-operator-controller-manager-5d89bdf8b7-wfkk6" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.679056 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq7tt\" (UniqueName: \"kubernetes.io/projected/b8be9f85-599f-4e47-99f5-c4e3d9ef27e9-kube-api-access-qq7tt\") pod \"metallb-operator-controller-manager-5d89bdf8b7-wfkk6\" (UID: \"b8be9f85-599f-4e47-99f5-c4e3d9ef27e9\") " pod="metallb-system/metallb-operator-controller-manager-5d89bdf8b7-wfkk6" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.679077 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b8be9f85-599f-4e47-99f5-c4e3d9ef27e9-apiservice-cert\") pod \"metallb-operator-controller-manager-5d89bdf8b7-wfkk6\" (UID: \"b8be9f85-599f-4e47-99f5-c4e3d9ef27e9\") " pod="metallb-system/metallb-operator-controller-manager-5d89bdf8b7-wfkk6" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.779920 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b8be9f85-599f-4e47-99f5-c4e3d9ef27e9-webhook-cert\") pod \"metallb-operator-controller-manager-5d89bdf8b7-wfkk6\" (UID: \"b8be9f85-599f-4e47-99f5-c4e3d9ef27e9\") " pod="metallb-system/metallb-operator-controller-manager-5d89bdf8b7-wfkk6" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.780342 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq7tt\" (UniqueName: \"kubernetes.io/projected/b8be9f85-599f-4e47-99f5-c4e3d9ef27e9-kube-api-access-qq7tt\") pod \"metallb-operator-controller-manager-5d89bdf8b7-wfkk6\" (UID: \"b8be9f85-599f-4e47-99f5-c4e3d9ef27e9\") " pod="metallb-system/metallb-operator-controller-manager-5d89bdf8b7-wfkk6" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.780470 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b8be9f85-599f-4e47-99f5-c4e3d9ef27e9-apiservice-cert\") pod \"metallb-operator-controller-manager-5d89bdf8b7-wfkk6\" (UID: \"b8be9f85-599f-4e47-99f5-c4e3d9ef27e9\") " pod="metallb-system/metallb-operator-controller-manager-5d89bdf8b7-wfkk6" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.786486 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b8be9f85-599f-4e47-99f5-c4e3d9ef27e9-apiservice-cert\") pod \"metallb-operator-controller-manager-5d89bdf8b7-wfkk6\" (UID: \"b8be9f85-599f-4e47-99f5-c4e3d9ef27e9\") " pod="metallb-system/metallb-operator-controller-manager-5d89bdf8b7-wfkk6" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.787413 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b8be9f85-599f-4e47-99f5-c4e3d9ef27e9-webhook-cert\") pod \"metallb-operator-controller-manager-5d89bdf8b7-wfkk6\" (UID: \"b8be9f85-599f-4e47-99f5-c4e3d9ef27e9\") " pod="metallb-system/metallb-operator-controller-manager-5d89bdf8b7-wfkk6" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.802936 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq7tt\" (UniqueName: \"kubernetes.io/projected/b8be9f85-599f-4e47-99f5-c4e3d9ef27e9-kube-api-access-qq7tt\") pod \"metallb-operator-controller-manager-5d89bdf8b7-wfkk6\" (UID: \"b8be9f85-599f-4e47-99f5-c4e3d9ef27e9\") " pod="metallb-system/metallb-operator-controller-manager-5d89bdf8b7-wfkk6" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.858398 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-648b974cdc-8fj4z"] Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.859258 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-648b974cdc-8fj4z" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.863554 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.863594 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.864093 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-9zjz9" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.876196 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-648b974cdc-8fj4z"] Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.881875 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/03782327-db08-4ac4-a753-2a23156d1329-webhook-cert\") pod \"metallb-operator-webhook-server-648b974cdc-8fj4z\" (UID: \"03782327-db08-4ac4-a753-2a23156d1329\") " pod="metallb-system/metallb-operator-webhook-server-648b974cdc-8fj4z" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.882124 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/03782327-db08-4ac4-a753-2a23156d1329-apiservice-cert\") pod \"metallb-operator-webhook-server-648b974cdc-8fj4z\" (UID: \"03782327-db08-4ac4-a753-2a23156d1329\") " pod="metallb-system/metallb-operator-webhook-server-648b974cdc-8fj4z" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.882320 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpgkv\" (UniqueName: \"kubernetes.io/projected/03782327-db08-4ac4-a753-2a23156d1329-kube-api-access-gpgkv\") pod \"metallb-operator-webhook-server-648b974cdc-8fj4z\" (UID: \"03782327-db08-4ac4-a753-2a23156d1329\") " pod="metallb-system/metallb-operator-webhook-server-648b974cdc-8fj4z" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.913038 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5d89bdf8b7-wfkk6" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.982794 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/03782327-db08-4ac4-a753-2a23156d1329-webhook-cert\") pod \"metallb-operator-webhook-server-648b974cdc-8fj4z\" (UID: \"03782327-db08-4ac4-a753-2a23156d1329\") " pod="metallb-system/metallb-operator-webhook-server-648b974cdc-8fj4z" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.983076 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/03782327-db08-4ac4-a753-2a23156d1329-apiservice-cert\") pod \"metallb-operator-webhook-server-648b974cdc-8fj4z\" (UID: \"03782327-db08-4ac4-a753-2a23156d1329\") " pod="metallb-system/metallb-operator-webhook-server-648b974cdc-8fj4z" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.983119 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpgkv\" (UniqueName: \"kubernetes.io/projected/03782327-db08-4ac4-a753-2a23156d1329-kube-api-access-gpgkv\") pod \"metallb-operator-webhook-server-648b974cdc-8fj4z\" (UID: \"03782327-db08-4ac4-a753-2a23156d1329\") " pod="metallb-system/metallb-operator-webhook-server-648b974cdc-8fj4z" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.988073 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/03782327-db08-4ac4-a753-2a23156d1329-webhook-cert\") pod \"metallb-operator-webhook-server-648b974cdc-8fj4z\" (UID: \"03782327-db08-4ac4-a753-2a23156d1329\") " pod="metallb-system/metallb-operator-webhook-server-648b974cdc-8fj4z" Oct 02 16:52:12 crc kubenswrapper[4808]: I1002 16:52:12.991188 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/03782327-db08-4ac4-a753-2a23156d1329-apiservice-cert\") pod \"metallb-operator-webhook-server-648b974cdc-8fj4z\" (UID: \"03782327-db08-4ac4-a753-2a23156d1329\") " pod="metallb-system/metallb-operator-webhook-server-648b974cdc-8fj4z" Oct 02 16:52:13 crc kubenswrapper[4808]: I1002 16:52:13.004914 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpgkv\" (UniqueName: \"kubernetes.io/projected/03782327-db08-4ac4-a753-2a23156d1329-kube-api-access-gpgkv\") pod \"metallb-operator-webhook-server-648b974cdc-8fj4z\" (UID: \"03782327-db08-4ac4-a753-2a23156d1329\") " pod="metallb-system/metallb-operator-webhook-server-648b974cdc-8fj4z" Oct 02 16:52:13 crc kubenswrapper[4808]: I1002 16:52:13.144412 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d89bdf8b7-wfkk6"] Oct 02 16:52:13 crc kubenswrapper[4808]: I1002 16:52:13.172501 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-648b974cdc-8fj4z" Oct 02 16:52:13 crc kubenswrapper[4808]: I1002 16:52:13.391928 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-648b974cdc-8fj4z"] Oct 02 16:52:13 crc kubenswrapper[4808]: W1002 16:52:13.402653 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03782327_db08_4ac4_a753_2a23156d1329.slice/crio-8baf6611bb0428bc8c95a901527d46d6fb87275cf74bc0efab05a85c9bdc538e WatchSource:0}: Error finding container 8baf6611bb0428bc8c95a901527d46d6fb87275cf74bc0efab05a85c9bdc538e: Status 404 returned error can't find the container with id 8baf6611bb0428bc8c95a901527d46d6fb87275cf74bc0efab05a85c9bdc538e Oct 02 16:52:13 crc kubenswrapper[4808]: I1002 16:52:13.880768 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5d89bdf8b7-wfkk6" event={"ID":"b8be9f85-599f-4e47-99f5-c4e3d9ef27e9","Type":"ContainerStarted","Data":"bd0cfc2e64d7ec269e8303551bdc342fddf026d356ba520fc2e40f0edfb1b8f6"} Oct 02 16:52:13 crc kubenswrapper[4808]: I1002 16:52:13.885086 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-648b974cdc-8fj4z" event={"ID":"03782327-db08-4ac4-a753-2a23156d1329","Type":"ContainerStarted","Data":"8baf6611bb0428bc8c95a901527d46d6fb87275cf74bc0efab05a85c9bdc538e"} Oct 02 16:52:18 crc kubenswrapper[4808]: I1002 16:52:18.916877 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5d89bdf8b7-wfkk6" event={"ID":"b8be9f85-599f-4e47-99f5-c4e3d9ef27e9","Type":"ContainerStarted","Data":"5727b5e99112f8a74abaae43f57d28d2a4866aa6c4daf084e8e018700384f51c"} Oct 02 16:52:18 crc kubenswrapper[4808]: I1002 16:52:18.917588 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5d89bdf8b7-wfkk6" Oct 02 16:52:18 crc kubenswrapper[4808]: I1002 16:52:18.918597 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-648b974cdc-8fj4z" event={"ID":"03782327-db08-4ac4-a753-2a23156d1329","Type":"ContainerStarted","Data":"bfb5af1fa3369bcbd2967fb93995c4dfda4f8264622d5c74b4faf80d12b5bbf2"} Oct 02 16:52:18 crc kubenswrapper[4808]: I1002 16:52:18.918831 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-648b974cdc-8fj4z" Oct 02 16:52:18 crc kubenswrapper[4808]: I1002 16:52:18.939986 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5d89bdf8b7-wfkk6" podStartSLOduration=1.5761893059999998 podStartE2EDuration="6.939964787s" podCreationTimestamp="2025-10-02 16:52:12 +0000 UTC" firstStartedPulling="2025-10-02 16:52:13.156515509 +0000 UTC m=+700.482044499" lastFinishedPulling="2025-10-02 16:52:18.52029097 +0000 UTC m=+705.845819980" observedRunningTime="2025-10-02 16:52:18.936940535 +0000 UTC m=+706.262469555" watchObservedRunningTime="2025-10-02 16:52:18.939964787 +0000 UTC m=+706.265493827" Oct 02 16:52:18 crc kubenswrapper[4808]: I1002 16:52:18.958862 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-648b974cdc-8fj4z" podStartSLOduration=1.814223829 podStartE2EDuration="6.958840909s" podCreationTimestamp="2025-10-02 16:52:12 +0000 UTC" firstStartedPulling="2025-10-02 16:52:13.407065971 +0000 UTC m=+700.732594971" lastFinishedPulling="2025-10-02 16:52:18.551683031 +0000 UTC m=+705.877212051" observedRunningTime="2025-10-02 16:52:18.958803858 +0000 UTC m=+706.284332908" watchObservedRunningTime="2025-10-02 16:52:18.958840909 +0000 UTC m=+706.284369909" Oct 02 16:52:33 crc kubenswrapper[4808]: I1002 16:52:33.177461 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-648b974cdc-8fj4z" Oct 02 16:52:52 crc kubenswrapper[4808]: I1002 16:52:52.917439 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5d89bdf8b7-wfkk6" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.604278 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-t9q4f"] Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.607201 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.608850 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-d5vzc"] Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.609324 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-7pmzd" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.609682 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d5vzc" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.610465 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.611608 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.615407 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.626372 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-d5vzc"] Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.651345 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-frr-startup\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.651410 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-frr-sockets\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.651480 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfvpj\" (UniqueName: \"kubernetes.io/projected/14914608-ea0e-4a8f-9bb7-88ff2aa55d12-kube-api-access-vfvpj\") pod \"frr-k8s-webhook-server-64bf5d555-d5vzc\" (UID: \"14914608-ea0e-4a8f-9bb7-88ff2aa55d12\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d5vzc" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.651517 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m45jd\" (UniqueName: \"kubernetes.io/projected/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-kube-api-access-m45jd\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.651627 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-reloader\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.651712 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14914608-ea0e-4a8f-9bb7-88ff2aa55d12-cert\") pod \"frr-k8s-webhook-server-64bf5d555-d5vzc\" (UID: \"14914608-ea0e-4a8f-9bb7-88ff2aa55d12\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d5vzc" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.651783 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-metrics-certs\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.651821 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-metrics\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.651850 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-frr-conf\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.716357 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-n9qkx"] Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.717473 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-n9qkx" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.719307 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.719524 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-klz4m" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.720687 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.721148 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.752787 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-metrics-certs\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: E1002 16:52:53.752964 4808 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.752985 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-metrics\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.753022 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-frr-conf\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: E1002 16:52:53.753040 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-metrics-certs podName:70e2d411-eb7c-46f9-bf6a-06ed29a0bd27 nodeName:}" failed. No retries permitted until 2025-10-02 16:52:54.2530224 +0000 UTC m=+741.578551400 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-metrics-certs") pod "frr-k8s-t9q4f" (UID: "70e2d411-eb7c-46f9-bf6a-06ed29a0bd27") : secret "frr-k8s-certs-secret" not found Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.753064 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/07aeac15-c264-4b20-ae80-d8d367224849-memberlist\") pod \"speaker-n9qkx\" (UID: \"07aeac15-c264-4b20-ae80-d8d367224849\") " pod="metallb-system/speaker-n9qkx" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.753102 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-frr-startup\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.753156 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-frr-sockets\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.753253 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/07aeac15-c264-4b20-ae80-d8d367224849-metrics-certs\") pod \"speaker-n9qkx\" (UID: \"07aeac15-c264-4b20-ae80-d8d367224849\") " pod="metallb-system/speaker-n9qkx" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.753280 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfvpj\" (UniqueName: \"kubernetes.io/projected/14914608-ea0e-4a8f-9bb7-88ff2aa55d12-kube-api-access-vfvpj\") pod \"frr-k8s-webhook-server-64bf5d555-d5vzc\" (UID: \"14914608-ea0e-4a8f-9bb7-88ff2aa55d12\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d5vzc" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.753299 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m45jd\" (UniqueName: \"kubernetes.io/projected/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-kube-api-access-m45jd\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.753317 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-reloader\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.753332 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7vdl\" (UniqueName: \"kubernetes.io/projected/07aeac15-c264-4b20-ae80-d8d367224849-kube-api-access-f7vdl\") pod \"speaker-n9qkx\" (UID: \"07aeac15-c264-4b20-ae80-d8d367224849\") " pod="metallb-system/speaker-n9qkx" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.753367 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14914608-ea0e-4a8f-9bb7-88ff2aa55d12-cert\") pod \"frr-k8s-webhook-server-64bf5d555-d5vzc\" (UID: \"14914608-ea0e-4a8f-9bb7-88ff2aa55d12\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d5vzc" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.753382 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/07aeac15-c264-4b20-ae80-d8d367224849-metallb-excludel2\") pod \"speaker-n9qkx\" (UID: \"07aeac15-c264-4b20-ae80-d8d367224849\") " pod="metallb-system/speaker-n9qkx" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.753454 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-frr-conf\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.753694 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-metrics\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.754473 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-frr-sockets\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.754575 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-frr-startup\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.754760 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-reloader\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.762729 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14914608-ea0e-4a8f-9bb7-88ff2aa55d12-cert\") pod \"frr-k8s-webhook-server-64bf5d555-d5vzc\" (UID: \"14914608-ea0e-4a8f-9bb7-88ff2aa55d12\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d5vzc" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.780315 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-57bdp"] Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.781442 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-57bdp" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.783557 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.791364 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfvpj\" (UniqueName: \"kubernetes.io/projected/14914608-ea0e-4a8f-9bb7-88ff2aa55d12-kube-api-access-vfvpj\") pod \"frr-k8s-webhook-server-64bf5d555-d5vzc\" (UID: \"14914608-ea0e-4a8f-9bb7-88ff2aa55d12\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d5vzc" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.792818 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m45jd\" (UniqueName: \"kubernetes.io/projected/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-kube-api-access-m45jd\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.807714 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-57bdp"] Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.854738 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/07aeac15-c264-4b20-ae80-d8d367224849-memberlist\") pod \"speaker-n9qkx\" (UID: \"07aeac15-c264-4b20-ae80-d8d367224849\") " pod="metallb-system/speaker-n9qkx" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.854802 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/575f788d-08ad-4138-82a1-fa7fa1543845-cert\") pod \"controller-68d546b9d8-57bdp\" (UID: \"575f788d-08ad-4138-82a1-fa7fa1543845\") " pod="metallb-system/controller-68d546b9d8-57bdp" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.854842 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/07aeac15-c264-4b20-ae80-d8d367224849-metrics-certs\") pod \"speaker-n9qkx\" (UID: \"07aeac15-c264-4b20-ae80-d8d367224849\") " pod="metallb-system/speaker-n9qkx" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.854865 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7vdl\" (UniqueName: \"kubernetes.io/projected/07aeac15-c264-4b20-ae80-d8d367224849-kube-api-access-f7vdl\") pod \"speaker-n9qkx\" (UID: \"07aeac15-c264-4b20-ae80-d8d367224849\") " pod="metallb-system/speaker-n9qkx" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.854889 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/575f788d-08ad-4138-82a1-fa7fa1543845-metrics-certs\") pod \"controller-68d546b9d8-57bdp\" (UID: \"575f788d-08ad-4138-82a1-fa7fa1543845\") " pod="metallb-system/controller-68d546b9d8-57bdp" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.854903 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/07aeac15-c264-4b20-ae80-d8d367224849-metallb-excludel2\") pod \"speaker-n9qkx\" (UID: \"07aeac15-c264-4b20-ae80-d8d367224849\") " pod="metallb-system/speaker-n9qkx" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.854929 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjlrd\" (UniqueName: \"kubernetes.io/projected/575f788d-08ad-4138-82a1-fa7fa1543845-kube-api-access-hjlrd\") pod \"controller-68d546b9d8-57bdp\" (UID: \"575f788d-08ad-4138-82a1-fa7fa1543845\") " pod="metallb-system/controller-68d546b9d8-57bdp" Oct 02 16:52:53 crc kubenswrapper[4808]: E1002 16:52:53.854895 4808 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 16:52:53 crc kubenswrapper[4808]: E1002 16:52:53.855317 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07aeac15-c264-4b20-ae80-d8d367224849-memberlist podName:07aeac15-c264-4b20-ae80-d8d367224849 nodeName:}" failed. No retries permitted until 2025-10-02 16:52:54.355294824 +0000 UTC m=+741.680823824 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/07aeac15-c264-4b20-ae80-d8d367224849-memberlist") pod "speaker-n9qkx" (UID: "07aeac15-c264-4b20-ae80-d8d367224849") : secret "metallb-memberlist" not found Oct 02 16:52:53 crc kubenswrapper[4808]: E1002 16:52:53.854959 4808 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Oct 02 16:52:53 crc kubenswrapper[4808]: E1002 16:52:53.855462 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07aeac15-c264-4b20-ae80-d8d367224849-metrics-certs podName:07aeac15-c264-4b20-ae80-d8d367224849 nodeName:}" failed. No retries permitted until 2025-10-02 16:52:54.355454209 +0000 UTC m=+741.680983209 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/07aeac15-c264-4b20-ae80-d8d367224849-metrics-certs") pod "speaker-n9qkx" (UID: "07aeac15-c264-4b20-ae80-d8d367224849") : secret "speaker-certs-secret" not found Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.855610 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/07aeac15-c264-4b20-ae80-d8d367224849-metallb-excludel2\") pod \"speaker-n9qkx\" (UID: \"07aeac15-c264-4b20-ae80-d8d367224849\") " pod="metallb-system/speaker-n9qkx" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.869475 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7vdl\" (UniqueName: \"kubernetes.io/projected/07aeac15-c264-4b20-ae80-d8d367224849-kube-api-access-f7vdl\") pod \"speaker-n9qkx\" (UID: \"07aeac15-c264-4b20-ae80-d8d367224849\") " pod="metallb-system/speaker-n9qkx" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.955844 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d5vzc" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.956245 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/575f788d-08ad-4138-82a1-fa7fa1543845-metrics-certs\") pod \"controller-68d546b9d8-57bdp\" (UID: \"575f788d-08ad-4138-82a1-fa7fa1543845\") " pod="metallb-system/controller-68d546b9d8-57bdp" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.956666 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjlrd\" (UniqueName: \"kubernetes.io/projected/575f788d-08ad-4138-82a1-fa7fa1543845-kube-api-access-hjlrd\") pod \"controller-68d546b9d8-57bdp\" (UID: \"575f788d-08ad-4138-82a1-fa7fa1543845\") " pod="metallb-system/controller-68d546b9d8-57bdp" Oct 02 16:52:53 crc kubenswrapper[4808]: E1002 16:52:53.956358 4808 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.956905 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/575f788d-08ad-4138-82a1-fa7fa1543845-cert\") pod \"controller-68d546b9d8-57bdp\" (UID: \"575f788d-08ad-4138-82a1-fa7fa1543845\") " pod="metallb-system/controller-68d546b9d8-57bdp" Oct 02 16:52:53 crc kubenswrapper[4808]: E1002 16:52:53.956937 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/575f788d-08ad-4138-82a1-fa7fa1543845-metrics-certs podName:575f788d-08ad-4138-82a1-fa7fa1543845 nodeName:}" failed. No retries permitted until 2025-10-02 16:52:54.456907481 +0000 UTC m=+741.782436481 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/575f788d-08ad-4138-82a1-fa7fa1543845-metrics-certs") pod "controller-68d546b9d8-57bdp" (UID: "575f788d-08ad-4138-82a1-fa7fa1543845") : secret "controller-certs-secret" not found Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.958199 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.972141 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/575f788d-08ad-4138-82a1-fa7fa1543845-cert\") pod \"controller-68d546b9d8-57bdp\" (UID: \"575f788d-08ad-4138-82a1-fa7fa1543845\") " pod="metallb-system/controller-68d546b9d8-57bdp" Oct 02 16:52:53 crc kubenswrapper[4808]: I1002 16:52:53.972791 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjlrd\" (UniqueName: \"kubernetes.io/projected/575f788d-08ad-4138-82a1-fa7fa1543845-kube-api-access-hjlrd\") pod \"controller-68d546b9d8-57bdp\" (UID: \"575f788d-08ad-4138-82a1-fa7fa1543845\") " pod="metallb-system/controller-68d546b9d8-57bdp" Oct 02 16:52:54 crc kubenswrapper[4808]: I1002 16:52:54.139074 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-d5vzc"] Oct 02 16:52:54 crc kubenswrapper[4808]: I1002 16:52:54.261365 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-metrics-certs\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:54 crc kubenswrapper[4808]: I1002 16:52:54.268584 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/70e2d411-eb7c-46f9-bf6a-06ed29a0bd27-metrics-certs\") pod \"frr-k8s-t9q4f\" (UID: \"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27\") " pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:54 crc kubenswrapper[4808]: I1002 16:52:54.362818 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/07aeac15-c264-4b20-ae80-d8d367224849-memberlist\") pod \"speaker-n9qkx\" (UID: \"07aeac15-c264-4b20-ae80-d8d367224849\") " pod="metallb-system/speaker-n9qkx" Oct 02 16:52:54 crc kubenswrapper[4808]: I1002 16:52:54.362896 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/07aeac15-c264-4b20-ae80-d8d367224849-metrics-certs\") pod \"speaker-n9qkx\" (UID: \"07aeac15-c264-4b20-ae80-d8d367224849\") " pod="metallb-system/speaker-n9qkx" Oct 02 16:52:54 crc kubenswrapper[4808]: E1002 16:52:54.363436 4808 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 16:52:54 crc kubenswrapper[4808]: E1002 16:52:54.363535 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07aeac15-c264-4b20-ae80-d8d367224849-memberlist podName:07aeac15-c264-4b20-ae80-d8d367224849 nodeName:}" failed. No retries permitted until 2025-10-02 16:52:55.363515203 +0000 UTC m=+742.689044203 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/07aeac15-c264-4b20-ae80-d8d367224849-memberlist") pod "speaker-n9qkx" (UID: "07aeac15-c264-4b20-ae80-d8d367224849") : secret "metallb-memberlist" not found Oct 02 16:52:54 crc kubenswrapper[4808]: I1002 16:52:54.365850 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/07aeac15-c264-4b20-ae80-d8d367224849-metrics-certs\") pod \"speaker-n9qkx\" (UID: \"07aeac15-c264-4b20-ae80-d8d367224849\") " pod="metallb-system/speaker-n9qkx" Oct 02 16:52:54 crc kubenswrapper[4808]: I1002 16:52:54.464295 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/575f788d-08ad-4138-82a1-fa7fa1543845-metrics-certs\") pod \"controller-68d546b9d8-57bdp\" (UID: \"575f788d-08ad-4138-82a1-fa7fa1543845\") " pod="metallb-system/controller-68d546b9d8-57bdp" Oct 02 16:52:54 crc kubenswrapper[4808]: I1002 16:52:54.467146 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/575f788d-08ad-4138-82a1-fa7fa1543845-metrics-certs\") pod \"controller-68d546b9d8-57bdp\" (UID: \"575f788d-08ad-4138-82a1-fa7fa1543845\") " pod="metallb-system/controller-68d546b9d8-57bdp" Oct 02 16:52:54 crc kubenswrapper[4808]: I1002 16:52:54.529933 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:52:54 crc kubenswrapper[4808]: I1002 16:52:54.741331 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-57bdp" Oct 02 16:52:55 crc kubenswrapper[4808]: I1002 16:52:55.152008 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d5vzc" event={"ID":"14914608-ea0e-4a8f-9bb7-88ff2aa55d12","Type":"ContainerStarted","Data":"34e6801d6d9a17e0560ad56d3e2dd400f53318fef800f983fa54dc2d65cb6207"} Oct 02 16:52:55 crc kubenswrapper[4808]: I1002 16:52:55.155700 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t9q4f" event={"ID":"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27","Type":"ContainerStarted","Data":"307332c40708bcbeb89beb3b51b5cca3c382d0772443111b52e6b47fecc86146"} Oct 02 16:52:55 crc kubenswrapper[4808]: I1002 16:52:55.180023 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-57bdp"] Oct 02 16:52:55 crc kubenswrapper[4808]: W1002 16:52:55.188114 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod575f788d_08ad_4138_82a1_fa7fa1543845.slice/crio-f569cdc834a43cefad8f3fb2f4f22d48d280be74c36d09c07cd8c92a24fee740 WatchSource:0}: Error finding container f569cdc834a43cefad8f3fb2f4f22d48d280be74c36d09c07cd8c92a24fee740: Status 404 returned error can't find the container with id f569cdc834a43cefad8f3fb2f4f22d48d280be74c36d09c07cd8c92a24fee740 Oct 02 16:52:55 crc kubenswrapper[4808]: I1002 16:52:55.378263 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/07aeac15-c264-4b20-ae80-d8d367224849-memberlist\") pod \"speaker-n9qkx\" (UID: \"07aeac15-c264-4b20-ae80-d8d367224849\") " pod="metallb-system/speaker-n9qkx" Oct 02 16:52:55 crc kubenswrapper[4808]: E1002 16:52:55.378530 4808 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 16:52:55 crc kubenswrapper[4808]: E1002 16:52:55.378693 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07aeac15-c264-4b20-ae80-d8d367224849-memberlist podName:07aeac15-c264-4b20-ae80-d8d367224849 nodeName:}" failed. No retries permitted until 2025-10-02 16:52:57.378626592 +0000 UTC m=+744.704155632 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/07aeac15-c264-4b20-ae80-d8d367224849-memberlist") pod "speaker-n9qkx" (UID: "07aeac15-c264-4b20-ae80-d8d367224849") : secret "metallb-memberlist" not found Oct 02 16:52:56 crc kubenswrapper[4808]: I1002 16:52:56.164847 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-57bdp" event={"ID":"575f788d-08ad-4138-82a1-fa7fa1543845","Type":"ContainerStarted","Data":"ee3e8e1c32e3bef86d8d3fdc6312a4df16d412495774e6fe6ef721b3c88b0e3a"} Oct 02 16:52:56 crc kubenswrapper[4808]: I1002 16:52:56.165115 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-57bdp" event={"ID":"575f788d-08ad-4138-82a1-fa7fa1543845","Type":"ContainerStarted","Data":"039e64a7b795ae4cfcabcce7bdb71ce5b3f590e344b88ab0da6b629699be9511"} Oct 02 16:52:56 crc kubenswrapper[4808]: I1002 16:52:56.165127 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-57bdp" event={"ID":"575f788d-08ad-4138-82a1-fa7fa1543845","Type":"ContainerStarted","Data":"f569cdc834a43cefad8f3fb2f4f22d48d280be74c36d09c07cd8c92a24fee740"} Oct 02 16:52:56 crc kubenswrapper[4808]: I1002 16:52:56.165942 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-57bdp" Oct 02 16:52:56 crc kubenswrapper[4808]: I1002 16:52:56.189896 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-57bdp" podStartSLOduration=3.189879482 podStartE2EDuration="3.189879482s" podCreationTimestamp="2025-10-02 16:52:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:52:56.189117381 +0000 UTC m=+743.514646381" watchObservedRunningTime="2025-10-02 16:52:56.189879482 +0000 UTC m=+743.515408482" Oct 02 16:52:57 crc kubenswrapper[4808]: I1002 16:52:57.406709 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/07aeac15-c264-4b20-ae80-d8d367224849-memberlist\") pod \"speaker-n9qkx\" (UID: \"07aeac15-c264-4b20-ae80-d8d367224849\") " pod="metallb-system/speaker-n9qkx" Oct 02 16:52:57 crc kubenswrapper[4808]: I1002 16:52:57.427877 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/07aeac15-c264-4b20-ae80-d8d367224849-memberlist\") pod \"speaker-n9qkx\" (UID: \"07aeac15-c264-4b20-ae80-d8d367224849\") " pod="metallb-system/speaker-n9qkx" Oct 02 16:52:57 crc kubenswrapper[4808]: I1002 16:52:57.701083 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-n9qkx" Oct 02 16:52:58 crc kubenswrapper[4808]: I1002 16:52:58.205203 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-n9qkx" event={"ID":"07aeac15-c264-4b20-ae80-d8d367224849","Type":"ContainerStarted","Data":"c30738d02d2d1077b744901bbfbcd49c6aad671a950fe5f35366c82831cb5753"} Oct 02 16:52:58 crc kubenswrapper[4808]: I1002 16:52:58.205562 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-n9qkx" event={"ID":"07aeac15-c264-4b20-ae80-d8d367224849","Type":"ContainerStarted","Data":"0e4659c8ae2c338aef58385d739ff36399021a9bc74356b8a86f80fbc526faac"} Oct 02 16:52:59 crc kubenswrapper[4808]: I1002 16:52:59.214099 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-n9qkx" event={"ID":"07aeac15-c264-4b20-ae80-d8d367224849","Type":"ContainerStarted","Data":"830888295b99aa2846fe807432b68c370a06f8d40e88c3fa37f788849d7db641"} Oct 02 16:52:59 crc kubenswrapper[4808]: I1002 16:52:59.214318 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-n9qkx" Oct 02 16:52:59 crc kubenswrapper[4808]: I1002 16:52:59.233080 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-n9qkx" podStartSLOduration=6.233059723 podStartE2EDuration="6.233059723s" podCreationTimestamp="2025-10-02 16:52:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:52:59.227778749 +0000 UTC m=+746.553307749" watchObservedRunningTime="2025-10-02 16:52:59.233059723 +0000 UTC m=+746.558588723" Oct 02 16:53:02 crc kubenswrapper[4808]: I1002 16:53:02.236194 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d5vzc" event={"ID":"14914608-ea0e-4a8f-9bb7-88ff2aa55d12","Type":"ContainerStarted","Data":"bd7a7336d32d4e5707d0dfc6a590a636ad3b6b9352541cb2426ffe6b6ee542ec"} Oct 02 16:53:02 crc kubenswrapper[4808]: I1002 16:53:02.236797 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d5vzc" Oct 02 16:53:02 crc kubenswrapper[4808]: I1002 16:53:02.239011 4808 generic.go:334] "Generic (PLEG): container finished" podID="70e2d411-eb7c-46f9-bf6a-06ed29a0bd27" containerID="5966da9aafb1c9ff3b52ed73c1fb67eeacd171d083de9087cd739b53ebd69655" exitCode=0 Oct 02 16:53:02 crc kubenswrapper[4808]: I1002 16:53:02.239080 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t9q4f" event={"ID":"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27","Type":"ContainerDied","Data":"5966da9aafb1c9ff3b52ed73c1fb67eeacd171d083de9087cd739b53ebd69655"} Oct 02 16:53:02 crc kubenswrapper[4808]: I1002 16:53:02.266870 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d5vzc" podStartSLOduration=1.609369322 podStartE2EDuration="9.266844759s" podCreationTimestamp="2025-10-02 16:52:53 +0000 UTC" firstStartedPulling="2025-10-02 16:52:54.154489531 +0000 UTC m=+741.480018521" lastFinishedPulling="2025-10-02 16:53:01.811964958 +0000 UTC m=+749.137493958" observedRunningTime="2025-10-02 16:53:02.257566815 +0000 UTC m=+749.583095855" watchObservedRunningTime="2025-10-02 16:53:02.266844759 +0000 UTC m=+749.592373789" Oct 02 16:53:03 crc kubenswrapper[4808]: I1002 16:53:03.250458 4808 generic.go:334] "Generic (PLEG): container finished" podID="70e2d411-eb7c-46f9-bf6a-06ed29a0bd27" containerID="057a5dff9ea59245514b9057efc5db2ef723d2a73360003ff1ae2f900bd2edb0" exitCode=0 Oct 02 16:53:03 crc kubenswrapper[4808]: I1002 16:53:03.250537 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t9q4f" event={"ID":"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27","Type":"ContainerDied","Data":"057a5dff9ea59245514b9057efc5db2ef723d2a73360003ff1ae2f900bd2edb0"} Oct 02 16:53:03 crc kubenswrapper[4808]: E1002 16:53:03.546874 4808 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70e2d411_eb7c_46f9_bf6a_06ed29a0bd27.slice/crio-5728763d36b06b2d14777f82eb7f5fb9e39bd20297514955696c50cbc9932fd3.scope\": RecentStats: unable to find data in memory cache]" Oct 02 16:53:04 crc kubenswrapper[4808]: I1002 16:53:04.265550 4808 generic.go:334] "Generic (PLEG): container finished" podID="70e2d411-eb7c-46f9-bf6a-06ed29a0bd27" containerID="5728763d36b06b2d14777f82eb7f5fb9e39bd20297514955696c50cbc9932fd3" exitCode=0 Oct 02 16:53:04 crc kubenswrapper[4808]: I1002 16:53:04.265601 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t9q4f" event={"ID":"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27","Type":"ContainerDied","Data":"5728763d36b06b2d14777f82eb7f5fb9e39bd20297514955696c50cbc9932fd3"} Oct 02 16:53:05 crc kubenswrapper[4808]: I1002 16:53:05.300306 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t9q4f" event={"ID":"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27","Type":"ContainerStarted","Data":"0f45a0bf87148fd3b87cec99db9033be82e1f384e2a9d625760baec3c64f4201"} Oct 02 16:53:05 crc kubenswrapper[4808]: I1002 16:53:05.300709 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t9q4f" event={"ID":"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27","Type":"ContainerStarted","Data":"90bf190bcbbda66947fcc6095164060adc139e45c235c24f8d72a617814eb04c"} Oct 02 16:53:05 crc kubenswrapper[4808]: I1002 16:53:05.300726 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t9q4f" event={"ID":"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27","Type":"ContainerStarted","Data":"11f9b139d5c726ff16ac2b442665f588518cc8075336231961401703edde5004"} Oct 02 16:53:05 crc kubenswrapper[4808]: I1002 16:53:05.300743 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t9q4f" event={"ID":"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27","Type":"ContainerStarted","Data":"24b711cfd2de187644c46ad9efca74d3412c7104884937bbb4dac15cdc3cf2ec"} Oct 02 16:53:05 crc kubenswrapper[4808]: I1002 16:53:05.300757 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t9q4f" event={"ID":"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27","Type":"ContainerStarted","Data":"169ecbbd85c8cb9989938c681b8e800e69b424f93fdeecebd5dd5d67277ed2b2"} Oct 02 16:53:06 crc kubenswrapper[4808]: I1002 16:53:06.310700 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t9q4f" event={"ID":"70e2d411-eb7c-46f9-bf6a-06ed29a0bd27","Type":"ContainerStarted","Data":"53ef5b9f6ec9c578dba56ff2f4e7a2a6b4ced8e91cccb25519ffeb4bd0f7af0e"} Oct 02 16:53:06 crc kubenswrapper[4808]: I1002 16:53:06.311133 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:53:06 crc kubenswrapper[4808]: I1002 16:53:06.347994 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-t9q4f" podStartSLOduration=6.221408915 podStartE2EDuration="13.347972404s" podCreationTimestamp="2025-10-02 16:52:53 +0000 UTC" firstStartedPulling="2025-10-02 16:52:54.662902154 +0000 UTC m=+741.988431164" lastFinishedPulling="2025-10-02 16:53:01.789465653 +0000 UTC m=+749.114994653" observedRunningTime="2025-10-02 16:53:06.343986685 +0000 UTC m=+753.669515725" watchObservedRunningTime="2025-10-02 16:53:06.347972404 +0000 UTC m=+753.673501434" Oct 02 16:53:07 crc kubenswrapper[4808]: I1002 16:53:07.864611 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qc629"] Oct 02 16:53:07 crc kubenswrapper[4808]: I1002 16:53:07.865012 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" podUID="c74723a8-81cf-4633-8655-e1460025f1b7" containerName="controller-manager" containerID="cri-o://8922f4f2f17ad396033076ece76e9ee8982c4f57182c3896a70dca03d9c2442a" gracePeriod=30 Oct 02 16:53:07 crc kubenswrapper[4808]: I1002 16:53:07.960306 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m"] Oct 02 16:53:07 crc kubenswrapper[4808]: I1002 16:53:07.960511 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" podUID="862243ee-b010-419a-884a-0824520ae00d" containerName="route-controller-manager" containerID="cri-o://b4ce5d6b3bca25999f28bf36169f5e81eb90dc05cf444ae9b91c0caf21df1c0c" gracePeriod=30 Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.241259 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.291086 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c74723a8-81cf-4633-8655-e1460025f1b7-proxy-ca-bundles\") pod \"c74723a8-81cf-4633-8655-e1460025f1b7\" (UID: \"c74723a8-81cf-4633-8655-e1460025f1b7\") " Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.291163 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kdgh\" (UniqueName: \"kubernetes.io/projected/c74723a8-81cf-4633-8655-e1460025f1b7-kube-api-access-5kdgh\") pod \"c74723a8-81cf-4633-8655-e1460025f1b7\" (UID: \"c74723a8-81cf-4633-8655-e1460025f1b7\") " Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.291222 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c74723a8-81cf-4633-8655-e1460025f1b7-serving-cert\") pod \"c74723a8-81cf-4633-8655-e1460025f1b7\" (UID: \"c74723a8-81cf-4633-8655-e1460025f1b7\") " Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.291287 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c74723a8-81cf-4633-8655-e1460025f1b7-client-ca\") pod \"c74723a8-81cf-4633-8655-e1460025f1b7\" (UID: \"c74723a8-81cf-4633-8655-e1460025f1b7\") " Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.291309 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c74723a8-81cf-4633-8655-e1460025f1b7-config\") pod \"c74723a8-81cf-4633-8655-e1460025f1b7\" (UID: \"c74723a8-81cf-4633-8655-e1460025f1b7\") " Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.291990 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c74723a8-81cf-4633-8655-e1460025f1b7-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "c74723a8-81cf-4633-8655-e1460025f1b7" (UID: "c74723a8-81cf-4633-8655-e1460025f1b7"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.292104 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c74723a8-81cf-4633-8655-e1460025f1b7-config" (OuterVolumeSpecName: "config") pod "c74723a8-81cf-4633-8655-e1460025f1b7" (UID: "c74723a8-81cf-4633-8655-e1460025f1b7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.292357 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c74723a8-81cf-4633-8655-e1460025f1b7-client-ca" (OuterVolumeSpecName: "client-ca") pod "c74723a8-81cf-4633-8655-e1460025f1b7" (UID: "c74723a8-81cf-4633-8655-e1460025f1b7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.301653 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c74723a8-81cf-4633-8655-e1460025f1b7-kube-api-access-5kdgh" (OuterVolumeSpecName: "kube-api-access-5kdgh") pod "c74723a8-81cf-4633-8655-e1460025f1b7" (UID: "c74723a8-81cf-4633-8655-e1460025f1b7"). InnerVolumeSpecName "kube-api-access-5kdgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.306428 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c74723a8-81cf-4633-8655-e1460025f1b7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c74723a8-81cf-4633-8655-e1460025f1b7" (UID: "c74723a8-81cf-4633-8655-e1460025f1b7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.331534 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.333659 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.333635 4808 generic.go:334] "Generic (PLEG): container finished" podID="c74723a8-81cf-4633-8655-e1460025f1b7" containerID="8922f4f2f17ad396033076ece76e9ee8982c4f57182c3896a70dca03d9c2442a" exitCode=0 Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.333741 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" event={"ID":"c74723a8-81cf-4633-8655-e1460025f1b7","Type":"ContainerDied","Data":"8922f4f2f17ad396033076ece76e9ee8982c4f57182c3896a70dca03d9c2442a"} Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.333800 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qc629" event={"ID":"c74723a8-81cf-4633-8655-e1460025f1b7","Type":"ContainerDied","Data":"4f4a104ef7859ab84dd7653986c84723fe3a1d0adcde8aaced7514f5f1d59a92"} Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.333819 4808 scope.go:117] "RemoveContainer" containerID="8922f4f2f17ad396033076ece76e9ee8982c4f57182c3896a70dca03d9c2442a" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.338112 4808 generic.go:334] "Generic (PLEG): container finished" podID="862243ee-b010-419a-884a-0824520ae00d" containerID="b4ce5d6b3bca25999f28bf36169f5e81eb90dc05cf444ae9b91c0caf21df1c0c" exitCode=0 Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.338139 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.338152 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" event={"ID":"862243ee-b010-419a-884a-0824520ae00d","Type":"ContainerDied","Data":"b4ce5d6b3bca25999f28bf36169f5e81eb90dc05cf444ae9b91c0caf21df1c0c"} Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.338179 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m" event={"ID":"862243ee-b010-419a-884a-0824520ae00d","Type":"ContainerDied","Data":"dacfc703a7e615cac96b4cc7b619998e61b36429a128a831bf2f1ca7b1942c91"} Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.361564 4808 scope.go:117] "RemoveContainer" containerID="8922f4f2f17ad396033076ece76e9ee8982c4f57182c3896a70dca03d9c2442a" Oct 02 16:53:08 crc kubenswrapper[4808]: E1002 16:53:08.365521 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8922f4f2f17ad396033076ece76e9ee8982c4f57182c3896a70dca03d9c2442a\": container with ID starting with 8922f4f2f17ad396033076ece76e9ee8982c4f57182c3896a70dca03d9c2442a not found: ID does not exist" containerID="8922f4f2f17ad396033076ece76e9ee8982c4f57182c3896a70dca03d9c2442a" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.365576 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8922f4f2f17ad396033076ece76e9ee8982c4f57182c3896a70dca03d9c2442a"} err="failed to get container status \"8922f4f2f17ad396033076ece76e9ee8982c4f57182c3896a70dca03d9c2442a\": rpc error: code = NotFound desc = could not find container \"8922f4f2f17ad396033076ece76e9ee8982c4f57182c3896a70dca03d9c2442a\": container with ID starting with 8922f4f2f17ad396033076ece76e9ee8982c4f57182c3896a70dca03d9c2442a not found: ID does not exist" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.365607 4808 scope.go:117] "RemoveContainer" containerID="b4ce5d6b3bca25999f28bf36169f5e81eb90dc05cf444ae9b91c0caf21df1c0c" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.372311 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qc629"] Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.377797 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qc629"] Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.392513 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/862243ee-b010-419a-884a-0824520ae00d-serving-cert\") pod \"862243ee-b010-419a-884a-0824520ae00d\" (UID: \"862243ee-b010-419a-884a-0824520ae00d\") " Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.392608 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/862243ee-b010-419a-884a-0824520ae00d-config\") pod \"862243ee-b010-419a-884a-0824520ae00d\" (UID: \"862243ee-b010-419a-884a-0824520ae00d\") " Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.392657 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/862243ee-b010-419a-884a-0824520ae00d-client-ca\") pod \"862243ee-b010-419a-884a-0824520ae00d\" (UID: \"862243ee-b010-419a-884a-0824520ae00d\") " Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.392711 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzq84\" (UniqueName: \"kubernetes.io/projected/862243ee-b010-419a-884a-0824520ae00d-kube-api-access-bzq84\") pod \"862243ee-b010-419a-884a-0824520ae00d\" (UID: \"862243ee-b010-419a-884a-0824520ae00d\") " Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.393048 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kdgh\" (UniqueName: \"kubernetes.io/projected/c74723a8-81cf-4633-8655-e1460025f1b7-kube-api-access-5kdgh\") on node \"crc\" DevicePath \"\"" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.393074 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c74723a8-81cf-4633-8655-e1460025f1b7-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.393092 4808 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c74723a8-81cf-4633-8655-e1460025f1b7-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.393103 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c74723a8-81cf-4633-8655-e1460025f1b7-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.393116 4808 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c74723a8-81cf-4633-8655-e1460025f1b7-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.394515 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/862243ee-b010-419a-884a-0824520ae00d-config" (OuterVolumeSpecName: "config") pod "862243ee-b010-419a-884a-0824520ae00d" (UID: "862243ee-b010-419a-884a-0824520ae00d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.395302 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/862243ee-b010-419a-884a-0824520ae00d-client-ca" (OuterVolumeSpecName: "client-ca") pod "862243ee-b010-419a-884a-0824520ae00d" (UID: "862243ee-b010-419a-884a-0824520ae00d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.397622 4808 scope.go:117] "RemoveContainer" containerID="b4ce5d6b3bca25999f28bf36169f5e81eb90dc05cf444ae9b91c0caf21df1c0c" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.397767 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/862243ee-b010-419a-884a-0824520ae00d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "862243ee-b010-419a-884a-0824520ae00d" (UID: "862243ee-b010-419a-884a-0824520ae00d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:53:08 crc kubenswrapper[4808]: E1002 16:53:08.398376 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4ce5d6b3bca25999f28bf36169f5e81eb90dc05cf444ae9b91c0caf21df1c0c\": container with ID starting with b4ce5d6b3bca25999f28bf36169f5e81eb90dc05cf444ae9b91c0caf21df1c0c not found: ID does not exist" containerID="b4ce5d6b3bca25999f28bf36169f5e81eb90dc05cf444ae9b91c0caf21df1c0c" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.398428 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4ce5d6b3bca25999f28bf36169f5e81eb90dc05cf444ae9b91c0caf21df1c0c"} err="failed to get container status \"b4ce5d6b3bca25999f28bf36169f5e81eb90dc05cf444ae9b91c0caf21df1c0c\": rpc error: code = NotFound desc = could not find container \"b4ce5d6b3bca25999f28bf36169f5e81eb90dc05cf444ae9b91c0caf21df1c0c\": container with ID starting with b4ce5d6b3bca25999f28bf36169f5e81eb90dc05cf444ae9b91c0caf21df1c0c not found: ID does not exist" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.405525 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/862243ee-b010-419a-884a-0824520ae00d-kube-api-access-bzq84" (OuterVolumeSpecName: "kube-api-access-bzq84") pod "862243ee-b010-419a-884a-0824520ae00d" (UID: "862243ee-b010-419a-884a-0824520ae00d"). InnerVolumeSpecName "kube-api-access-bzq84". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.494990 4808 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/862243ee-b010-419a-884a-0824520ae00d-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.495042 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzq84\" (UniqueName: \"kubernetes.io/projected/862243ee-b010-419a-884a-0824520ae00d-kube-api-access-bzq84\") on node \"crc\" DevicePath \"\"" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.495061 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/862243ee-b010-419a-884a-0824520ae00d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.495080 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/862243ee-b010-419a-884a-0824520ae00d-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.682774 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m"] Oct 02 16:53:08 crc kubenswrapper[4808]: I1002 16:53:08.687171 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6br9m"] Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.405327 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="862243ee-b010-419a-884a-0824520ae00d" path="/var/lib/kubelet/pods/862243ee-b010-419a-884a-0824520ae00d/volumes" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.406327 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c74723a8-81cf-4633-8655-e1460025f1b7" path="/var/lib/kubelet/pods/c74723a8-81cf-4633-8655-e1460025f1b7/volumes" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.530934 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.581305 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.892156 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q"] Oct 02 16:53:09 crc kubenswrapper[4808]: E1002 16:53:09.892580 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c74723a8-81cf-4633-8655-e1460025f1b7" containerName="controller-manager" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.892640 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c74723a8-81cf-4633-8655-e1460025f1b7" containerName="controller-manager" Oct 02 16:53:09 crc kubenswrapper[4808]: E1002 16:53:09.892731 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862243ee-b010-419a-884a-0824520ae00d" containerName="route-controller-manager" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.892782 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="862243ee-b010-419a-884a-0824520ae00d" containerName="route-controller-manager" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.892916 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="862243ee-b010-419a-884a-0824520ae00d" containerName="route-controller-manager" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.892982 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c74723a8-81cf-4633-8655-e1460025f1b7" containerName="controller-manager" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.893384 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.896682 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.896726 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.898496 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.898673 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.898860 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.899041 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.907016 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q"] Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.913387 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs"] Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.914279 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.916622 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.916977 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.917111 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.917271 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.917397 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.917655 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.923265 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 02 16:53:09 crc kubenswrapper[4808]: I1002 16:53:09.937144 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs"] Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.014395 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7ac4ef1-9bed-4416-b700-4e48c91c0391-serving-cert\") pod \"controller-manager-6bc44bf7cb-lgsfs\" (UID: \"e7ac4ef1-9bed-4416-b700-4e48c91c0391\") " pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.014445 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71fabec9-6b2c-4e31-aef3-1d63299ceec3-client-ca\") pod \"route-controller-manager-c5b56b4bf-v442q\" (UID: \"71fabec9-6b2c-4e31-aef3-1d63299ceec3\") " pod="openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.014466 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl9vq\" (UniqueName: \"kubernetes.io/projected/71fabec9-6b2c-4e31-aef3-1d63299ceec3-kube-api-access-vl9vq\") pod \"route-controller-manager-c5b56b4bf-v442q\" (UID: \"71fabec9-6b2c-4e31-aef3-1d63299ceec3\") " pod="openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.014490 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz5pt\" (UniqueName: \"kubernetes.io/projected/e7ac4ef1-9bed-4416-b700-4e48c91c0391-kube-api-access-nz5pt\") pod \"controller-manager-6bc44bf7cb-lgsfs\" (UID: \"e7ac4ef1-9bed-4416-b700-4e48c91c0391\") " pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.014560 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e7ac4ef1-9bed-4416-b700-4e48c91c0391-client-ca\") pod \"controller-manager-6bc44bf7cb-lgsfs\" (UID: \"e7ac4ef1-9bed-4416-b700-4e48c91c0391\") " pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.014580 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e7ac4ef1-9bed-4416-b700-4e48c91c0391-proxy-ca-bundles\") pod \"controller-manager-6bc44bf7cb-lgsfs\" (UID: \"e7ac4ef1-9bed-4416-b700-4e48c91c0391\") " pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.014595 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71fabec9-6b2c-4e31-aef3-1d63299ceec3-config\") pod \"route-controller-manager-c5b56b4bf-v442q\" (UID: \"71fabec9-6b2c-4e31-aef3-1d63299ceec3\") " pod="openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.014612 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7ac4ef1-9bed-4416-b700-4e48c91c0391-config\") pod \"controller-manager-6bc44bf7cb-lgsfs\" (UID: \"e7ac4ef1-9bed-4416-b700-4e48c91c0391\") " pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.014628 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71fabec9-6b2c-4e31-aef3-1d63299ceec3-serving-cert\") pod \"route-controller-manager-c5b56b4bf-v442q\" (UID: \"71fabec9-6b2c-4e31-aef3-1d63299ceec3\") " pod="openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.118311 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e7ac4ef1-9bed-4416-b700-4e48c91c0391-client-ca\") pod \"controller-manager-6bc44bf7cb-lgsfs\" (UID: \"e7ac4ef1-9bed-4416-b700-4e48c91c0391\") " pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.118363 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e7ac4ef1-9bed-4416-b700-4e48c91c0391-proxy-ca-bundles\") pod \"controller-manager-6bc44bf7cb-lgsfs\" (UID: \"e7ac4ef1-9bed-4416-b700-4e48c91c0391\") " pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.118387 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71fabec9-6b2c-4e31-aef3-1d63299ceec3-config\") pod \"route-controller-manager-c5b56b4bf-v442q\" (UID: \"71fabec9-6b2c-4e31-aef3-1d63299ceec3\") " pod="openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.118407 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7ac4ef1-9bed-4416-b700-4e48c91c0391-config\") pod \"controller-manager-6bc44bf7cb-lgsfs\" (UID: \"e7ac4ef1-9bed-4416-b700-4e48c91c0391\") " pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.120342 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71fabec9-6b2c-4e31-aef3-1d63299ceec3-serving-cert\") pod \"route-controller-manager-c5b56b4bf-v442q\" (UID: \"71fabec9-6b2c-4e31-aef3-1d63299ceec3\") " pod="openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.120419 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7ac4ef1-9bed-4416-b700-4e48c91c0391-serving-cert\") pod \"controller-manager-6bc44bf7cb-lgsfs\" (UID: \"e7ac4ef1-9bed-4416-b700-4e48c91c0391\") " pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.120453 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71fabec9-6b2c-4e31-aef3-1d63299ceec3-client-ca\") pod \"route-controller-manager-c5b56b4bf-v442q\" (UID: \"71fabec9-6b2c-4e31-aef3-1d63299ceec3\") " pod="openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.119774 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7ac4ef1-9bed-4416-b700-4e48c91c0391-config\") pod \"controller-manager-6bc44bf7cb-lgsfs\" (UID: \"e7ac4ef1-9bed-4416-b700-4e48c91c0391\") " pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.120476 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl9vq\" (UniqueName: \"kubernetes.io/projected/71fabec9-6b2c-4e31-aef3-1d63299ceec3-kube-api-access-vl9vq\") pod \"route-controller-manager-c5b56b4bf-v442q\" (UID: \"71fabec9-6b2c-4e31-aef3-1d63299ceec3\") " pod="openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.120562 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz5pt\" (UniqueName: \"kubernetes.io/projected/e7ac4ef1-9bed-4416-b700-4e48c91c0391-kube-api-access-nz5pt\") pod \"controller-manager-6bc44bf7cb-lgsfs\" (UID: \"e7ac4ef1-9bed-4416-b700-4e48c91c0391\") " pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.119438 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e7ac4ef1-9bed-4416-b700-4e48c91c0391-client-ca\") pod \"controller-manager-6bc44bf7cb-lgsfs\" (UID: \"e7ac4ef1-9bed-4416-b700-4e48c91c0391\") " pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.120339 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e7ac4ef1-9bed-4416-b700-4e48c91c0391-proxy-ca-bundles\") pod \"controller-manager-6bc44bf7cb-lgsfs\" (UID: \"e7ac4ef1-9bed-4416-b700-4e48c91c0391\") " pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.121139 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71fabec9-6b2c-4e31-aef3-1d63299ceec3-client-ca\") pod \"route-controller-manager-c5b56b4bf-v442q\" (UID: \"71fabec9-6b2c-4e31-aef3-1d63299ceec3\") " pod="openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.121390 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71fabec9-6b2c-4e31-aef3-1d63299ceec3-config\") pod \"route-controller-manager-c5b56b4bf-v442q\" (UID: \"71fabec9-6b2c-4e31-aef3-1d63299ceec3\") " pod="openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.125684 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7ac4ef1-9bed-4416-b700-4e48c91c0391-serving-cert\") pod \"controller-manager-6bc44bf7cb-lgsfs\" (UID: \"e7ac4ef1-9bed-4416-b700-4e48c91c0391\") " pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.128770 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71fabec9-6b2c-4e31-aef3-1d63299ceec3-serving-cert\") pod \"route-controller-manager-c5b56b4bf-v442q\" (UID: \"71fabec9-6b2c-4e31-aef3-1d63299ceec3\") " pod="openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.135743 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl9vq\" (UniqueName: \"kubernetes.io/projected/71fabec9-6b2c-4e31-aef3-1d63299ceec3-kube-api-access-vl9vq\") pod \"route-controller-manager-c5b56b4bf-v442q\" (UID: \"71fabec9-6b2c-4e31-aef3-1d63299ceec3\") " pod="openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.138148 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz5pt\" (UniqueName: \"kubernetes.io/projected/e7ac4ef1-9bed-4416-b700-4e48c91c0391-kube-api-access-nz5pt\") pod \"controller-manager-6bc44bf7cb-lgsfs\" (UID: \"e7ac4ef1-9bed-4416-b700-4e48c91c0391\") " pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.218718 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.243763 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.470307 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs"] Oct 02 16:53:10 crc kubenswrapper[4808]: W1002 16:53:10.475031 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7ac4ef1_9bed_4416_b700_4e48c91c0391.slice/crio-afd2e4b08ed52b7c88aeb1c002d8a6a8077e5cb6d3ca0a38f12b03795ad47458 WatchSource:0}: Error finding container afd2e4b08ed52b7c88aeb1c002d8a6a8077e5cb6d3ca0a38f12b03795ad47458: Status 404 returned error can't find the container with id afd2e4b08ed52b7c88aeb1c002d8a6a8077e5cb6d3ca0a38f12b03795ad47458 Oct 02 16:53:10 crc kubenswrapper[4808]: I1002 16:53:10.634339 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q"] Oct 02 16:53:10 crc kubenswrapper[4808]: W1002 16:53:10.639287 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71fabec9_6b2c_4e31_aef3_1d63299ceec3.slice/crio-ea1a58ad6c90be71c48d18d11f24e6c673ef73e46a76bc97405254606eab406e WatchSource:0}: Error finding container ea1a58ad6c90be71c48d18d11f24e6c673ef73e46a76bc97405254606eab406e: Status 404 returned error can't find the container with id ea1a58ad6c90be71c48d18d11f24e6c673ef73e46a76bc97405254606eab406e Oct 02 16:53:11 crc kubenswrapper[4808]: I1002 16:53:11.363489 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q" event={"ID":"71fabec9-6b2c-4e31-aef3-1d63299ceec3","Type":"ContainerStarted","Data":"f090926c37e194fdb0069f52b672ae718b449ee37faa0194bae26465140f38f7"} Oct 02 16:53:11 crc kubenswrapper[4808]: I1002 16:53:11.363807 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q" Oct 02 16:53:11 crc kubenswrapper[4808]: I1002 16:53:11.363822 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q" event={"ID":"71fabec9-6b2c-4e31-aef3-1d63299ceec3","Type":"ContainerStarted","Data":"ea1a58ad6c90be71c48d18d11f24e6c673ef73e46a76bc97405254606eab406e"} Oct 02 16:53:11 crc kubenswrapper[4808]: I1002 16:53:11.364642 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" event={"ID":"e7ac4ef1-9bed-4416-b700-4e48c91c0391","Type":"ContainerStarted","Data":"0b14d902164ce9d6b46f50b6919855a0fb2cd316074ba39f33410b05e912c2f1"} Oct 02 16:53:11 crc kubenswrapper[4808]: I1002 16:53:11.364680 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" event={"ID":"e7ac4ef1-9bed-4416-b700-4e48c91c0391","Type":"ContainerStarted","Data":"afd2e4b08ed52b7c88aeb1c002d8a6a8077e5cb6d3ca0a38f12b03795ad47458"} Oct 02 16:53:11 crc kubenswrapper[4808]: I1002 16:53:11.364838 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" Oct 02 16:53:11 crc kubenswrapper[4808]: I1002 16:53:11.370373 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" Oct 02 16:53:11 crc kubenswrapper[4808]: I1002 16:53:11.372138 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q" Oct 02 16:53:11 crc kubenswrapper[4808]: I1002 16:53:11.382558 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-c5b56b4bf-v442q" podStartSLOduration=2.3825333349999998 podStartE2EDuration="2.382533335s" podCreationTimestamp="2025-10-02 16:53:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:53:11.378426642 +0000 UTC m=+758.703955642" watchObservedRunningTime="2025-10-02 16:53:11.382533335 +0000 UTC m=+758.708062335" Oct 02 16:53:11 crc kubenswrapper[4808]: I1002 16:53:11.393979 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6bc44bf7cb-lgsfs" podStartSLOduration=3.3939579970000002 podStartE2EDuration="3.393957997s" podCreationTimestamp="2025-10-02 16:53:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:53:11.393059972 +0000 UTC m=+758.718588972" watchObservedRunningTime="2025-10-02 16:53:11.393957997 +0000 UTC m=+758.719486997" Oct 02 16:53:13 crc kubenswrapper[4808]: I1002 16:53:13.962767 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d5vzc" Oct 02 16:53:14 crc kubenswrapper[4808]: I1002 16:53:14.533812 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-t9q4f" Oct 02 16:53:14 crc kubenswrapper[4808]: I1002 16:53:14.749059 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-57bdp" Oct 02 16:53:17 crc kubenswrapper[4808]: I1002 16:53:17.707026 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-n9qkx" Oct 02 16:53:18 crc kubenswrapper[4808]: I1002 16:53:18.664976 4808 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 16:53:19 crc kubenswrapper[4808]: I1002 16:53:19.281833 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2kqnl"] Oct 02 16:53:19 crc kubenswrapper[4808]: I1002 16:53:19.284442 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2kqnl" Oct 02 16:53:19 crc kubenswrapper[4808]: I1002 16:53:19.297791 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2kqnl"] Oct 02 16:53:19 crc kubenswrapper[4808]: I1002 16:53:19.353221 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d20c4266-42c2-49d2-9fb0-323f5fd208b7-utilities\") pod \"redhat-operators-2kqnl\" (UID: \"d20c4266-42c2-49d2-9fb0-323f5fd208b7\") " pod="openshift-marketplace/redhat-operators-2kqnl" Oct 02 16:53:19 crc kubenswrapper[4808]: I1002 16:53:19.353717 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7mxx\" (UniqueName: \"kubernetes.io/projected/d20c4266-42c2-49d2-9fb0-323f5fd208b7-kube-api-access-s7mxx\") pod \"redhat-operators-2kqnl\" (UID: \"d20c4266-42c2-49d2-9fb0-323f5fd208b7\") " pod="openshift-marketplace/redhat-operators-2kqnl" Oct 02 16:53:19 crc kubenswrapper[4808]: I1002 16:53:19.353748 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d20c4266-42c2-49d2-9fb0-323f5fd208b7-catalog-content\") pod \"redhat-operators-2kqnl\" (UID: \"d20c4266-42c2-49d2-9fb0-323f5fd208b7\") " pod="openshift-marketplace/redhat-operators-2kqnl" Oct 02 16:53:19 crc kubenswrapper[4808]: I1002 16:53:19.454701 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d20c4266-42c2-49d2-9fb0-323f5fd208b7-utilities\") pod \"redhat-operators-2kqnl\" (UID: \"d20c4266-42c2-49d2-9fb0-323f5fd208b7\") " pod="openshift-marketplace/redhat-operators-2kqnl" Oct 02 16:53:19 crc kubenswrapper[4808]: I1002 16:53:19.454783 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7mxx\" (UniqueName: \"kubernetes.io/projected/d20c4266-42c2-49d2-9fb0-323f5fd208b7-kube-api-access-s7mxx\") pod \"redhat-operators-2kqnl\" (UID: \"d20c4266-42c2-49d2-9fb0-323f5fd208b7\") " pod="openshift-marketplace/redhat-operators-2kqnl" Oct 02 16:53:19 crc kubenswrapper[4808]: I1002 16:53:19.454842 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d20c4266-42c2-49d2-9fb0-323f5fd208b7-catalog-content\") pod \"redhat-operators-2kqnl\" (UID: \"d20c4266-42c2-49d2-9fb0-323f5fd208b7\") " pod="openshift-marketplace/redhat-operators-2kqnl" Oct 02 16:53:19 crc kubenswrapper[4808]: I1002 16:53:19.455998 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d20c4266-42c2-49d2-9fb0-323f5fd208b7-utilities\") pod \"redhat-operators-2kqnl\" (UID: \"d20c4266-42c2-49d2-9fb0-323f5fd208b7\") " pod="openshift-marketplace/redhat-operators-2kqnl" Oct 02 16:53:19 crc kubenswrapper[4808]: I1002 16:53:19.456053 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d20c4266-42c2-49d2-9fb0-323f5fd208b7-catalog-content\") pod \"redhat-operators-2kqnl\" (UID: \"d20c4266-42c2-49d2-9fb0-323f5fd208b7\") " pod="openshift-marketplace/redhat-operators-2kqnl" Oct 02 16:53:19 crc kubenswrapper[4808]: I1002 16:53:19.482445 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7mxx\" (UniqueName: \"kubernetes.io/projected/d20c4266-42c2-49d2-9fb0-323f5fd208b7-kube-api-access-s7mxx\") pod \"redhat-operators-2kqnl\" (UID: \"d20c4266-42c2-49d2-9fb0-323f5fd208b7\") " pod="openshift-marketplace/redhat-operators-2kqnl" Oct 02 16:53:19 crc kubenswrapper[4808]: I1002 16:53:19.620005 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2kqnl" Oct 02 16:53:20 crc kubenswrapper[4808]: I1002 16:53:20.033907 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2kqnl"] Oct 02 16:53:20 crc kubenswrapper[4808]: W1002 16:53:20.041854 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd20c4266_42c2_49d2_9fb0_323f5fd208b7.slice/crio-1d4d49713581ec695c8efaa1e743f4843e05edad47932676bfe2202c90c16af3 WatchSource:0}: Error finding container 1d4d49713581ec695c8efaa1e743f4843e05edad47932676bfe2202c90c16af3: Status 404 returned error can't find the container with id 1d4d49713581ec695c8efaa1e743f4843e05edad47932676bfe2202c90c16af3 Oct 02 16:53:20 crc kubenswrapper[4808]: I1002 16:53:20.423696 4808 generic.go:334] "Generic (PLEG): container finished" podID="d20c4266-42c2-49d2-9fb0-323f5fd208b7" containerID="2fb106fc6c5d3d1caf8438a9aebebd14ddd789904f3441292b27c8ad53f0815d" exitCode=0 Oct 02 16:53:20 crc kubenswrapper[4808]: I1002 16:53:20.423738 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2kqnl" event={"ID":"d20c4266-42c2-49d2-9fb0-323f5fd208b7","Type":"ContainerDied","Data":"2fb106fc6c5d3d1caf8438a9aebebd14ddd789904f3441292b27c8ad53f0815d"} Oct 02 16:53:20 crc kubenswrapper[4808]: I1002 16:53:20.423765 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2kqnl" event={"ID":"d20c4266-42c2-49d2-9fb0-323f5fd208b7","Type":"ContainerStarted","Data":"1d4d49713581ec695c8efaa1e743f4843e05edad47932676bfe2202c90c16af3"} Oct 02 16:53:22 crc kubenswrapper[4808]: I1002 16:53:22.438627 4808 generic.go:334] "Generic (PLEG): container finished" podID="d20c4266-42c2-49d2-9fb0-323f5fd208b7" containerID="fc05e3fce5d6cbd9e6c87112ec4f814d919179e6922f1371e266bad6cc21f707" exitCode=0 Oct 02 16:53:22 crc kubenswrapper[4808]: I1002 16:53:22.438838 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2kqnl" event={"ID":"d20c4266-42c2-49d2-9fb0-323f5fd208b7","Type":"ContainerDied","Data":"fc05e3fce5d6cbd9e6c87112ec4f814d919179e6922f1371e266bad6cc21f707"} Oct 02 16:53:23 crc kubenswrapper[4808]: I1002 16:53:23.448189 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2kqnl" event={"ID":"d20c4266-42c2-49d2-9fb0-323f5fd208b7","Type":"ContainerStarted","Data":"bd334d1eb16c3070282a007c8afd10b65e6b0081d4287ceedbb4af46f72dcaf3"} Oct 02 16:53:23 crc kubenswrapper[4808]: I1002 16:53:23.474327 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2kqnl" podStartSLOduration=1.950989724 podStartE2EDuration="4.47429726s" podCreationTimestamp="2025-10-02 16:53:19 +0000 UTC" firstStartedPulling="2025-10-02 16:53:20.425600357 +0000 UTC m=+767.751129357" lastFinishedPulling="2025-10-02 16:53:22.948907853 +0000 UTC m=+770.274436893" observedRunningTime="2025-10-02 16:53:23.46989493 +0000 UTC m=+770.795423970" watchObservedRunningTime="2025-10-02 16:53:23.47429726 +0000 UTC m=+770.799826290" Oct 02 16:53:23 crc kubenswrapper[4808]: I1002 16:53:23.651430 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-7p6jd"] Oct 02 16:53:23 crc kubenswrapper[4808]: I1002 16:53:23.652525 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7p6jd" Oct 02 16:53:23 crc kubenswrapper[4808]: I1002 16:53:23.657645 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 02 16:53:23 crc kubenswrapper[4808]: I1002 16:53:23.657800 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-ck9xh" Oct 02 16:53:23 crc kubenswrapper[4808]: I1002 16:53:23.657997 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 02 16:53:23 crc kubenswrapper[4808]: I1002 16:53:23.674619 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7p6jd"] Oct 02 16:53:23 crc kubenswrapper[4808]: I1002 16:53:23.726583 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zm6h\" (UniqueName: \"kubernetes.io/projected/d02c0e95-ce3f-42d1-9fb9-9ff040b18447-kube-api-access-7zm6h\") pod \"openstack-operator-index-7p6jd\" (UID: \"d02c0e95-ce3f-42d1-9fb9-9ff040b18447\") " pod="openstack-operators/openstack-operator-index-7p6jd" Oct 02 16:53:23 crc kubenswrapper[4808]: I1002 16:53:23.840343 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zm6h\" (UniqueName: \"kubernetes.io/projected/d02c0e95-ce3f-42d1-9fb9-9ff040b18447-kube-api-access-7zm6h\") pod \"openstack-operator-index-7p6jd\" (UID: \"d02c0e95-ce3f-42d1-9fb9-9ff040b18447\") " pod="openstack-operators/openstack-operator-index-7p6jd" Oct 02 16:53:23 crc kubenswrapper[4808]: I1002 16:53:23.861683 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zm6h\" (UniqueName: \"kubernetes.io/projected/d02c0e95-ce3f-42d1-9fb9-9ff040b18447-kube-api-access-7zm6h\") pod \"openstack-operator-index-7p6jd\" (UID: \"d02c0e95-ce3f-42d1-9fb9-9ff040b18447\") " pod="openstack-operators/openstack-operator-index-7p6jd" Oct 02 16:53:23 crc kubenswrapper[4808]: I1002 16:53:23.970345 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7p6jd" Oct 02 16:53:24 crc kubenswrapper[4808]: I1002 16:53:24.451494 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7p6jd"] Oct 02 16:53:25 crc kubenswrapper[4808]: I1002 16:53:25.465904 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7p6jd" event={"ID":"d02c0e95-ce3f-42d1-9fb9-9ff040b18447","Type":"ContainerStarted","Data":"1eeaab89c6fa053096dac6de33882029314e5c618043e9509255bf5f386c47a6"} Oct 02 16:53:27 crc kubenswrapper[4808]: I1002 16:53:27.497862 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7p6jd" event={"ID":"d02c0e95-ce3f-42d1-9fb9-9ff040b18447","Type":"ContainerStarted","Data":"bd0f92e2bdf532d80ff0baf30c872ef9d4fc1827714aff0feaac0034c67c17da"} Oct 02 16:53:27 crc kubenswrapper[4808]: I1002 16:53:27.518547 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-7p6jd" podStartSLOduration=1.8785523130000001 podStartE2EDuration="4.518527707s" podCreationTimestamp="2025-10-02 16:53:23 +0000 UTC" firstStartedPulling="2025-10-02 16:53:24.462609467 +0000 UTC m=+771.788138467" lastFinishedPulling="2025-10-02 16:53:27.102584821 +0000 UTC m=+774.428113861" observedRunningTime="2025-10-02 16:53:27.513898751 +0000 UTC m=+774.839427771" watchObservedRunningTime="2025-10-02 16:53:27.518527707 +0000 UTC m=+774.844056697" Oct 02 16:53:27 crc kubenswrapper[4808]: I1002 16:53:27.839628 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-7p6jd"] Oct 02 16:53:28 crc kubenswrapper[4808]: I1002 16:53:28.447483 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-zstrp"] Oct 02 16:53:28 crc kubenswrapper[4808]: I1002 16:53:28.448469 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zstrp" Oct 02 16:53:28 crc kubenswrapper[4808]: I1002 16:53:28.463808 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zstrp"] Oct 02 16:53:28 crc kubenswrapper[4808]: I1002 16:53:28.619703 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjqls\" (UniqueName: \"kubernetes.io/projected/1883774e-2d09-400d-95bd-490deded815b-kube-api-access-xjqls\") pod \"openstack-operator-index-zstrp\" (UID: \"1883774e-2d09-400d-95bd-490deded815b\") " pod="openstack-operators/openstack-operator-index-zstrp" Oct 02 16:53:28 crc kubenswrapper[4808]: I1002 16:53:28.721189 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjqls\" (UniqueName: \"kubernetes.io/projected/1883774e-2d09-400d-95bd-490deded815b-kube-api-access-xjqls\") pod \"openstack-operator-index-zstrp\" (UID: \"1883774e-2d09-400d-95bd-490deded815b\") " pod="openstack-operators/openstack-operator-index-zstrp" Oct 02 16:53:28 crc kubenswrapper[4808]: I1002 16:53:28.756796 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjqls\" (UniqueName: \"kubernetes.io/projected/1883774e-2d09-400d-95bd-490deded815b-kube-api-access-xjqls\") pod \"openstack-operator-index-zstrp\" (UID: \"1883774e-2d09-400d-95bd-490deded815b\") " pod="openstack-operators/openstack-operator-index-zstrp" Oct 02 16:53:28 crc kubenswrapper[4808]: I1002 16:53:28.783953 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zstrp" Oct 02 16:53:29 crc kubenswrapper[4808]: I1002 16:53:29.268577 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zstrp"] Oct 02 16:53:29 crc kubenswrapper[4808]: W1002 16:53:29.281491 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1883774e_2d09_400d_95bd_490deded815b.slice/crio-b4b69c19725ad9cfedfadf02e0f76a32f2998473a4fda2e138fe4e4c67c875e8 WatchSource:0}: Error finding container b4b69c19725ad9cfedfadf02e0f76a32f2998473a4fda2e138fe4e4c67c875e8: Status 404 returned error can't find the container with id b4b69c19725ad9cfedfadf02e0f76a32f2998473a4fda2e138fe4e4c67c875e8 Oct 02 16:53:29 crc kubenswrapper[4808]: I1002 16:53:29.515575 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zstrp" event={"ID":"1883774e-2d09-400d-95bd-490deded815b","Type":"ContainerStarted","Data":"b4b69c19725ad9cfedfadf02e0f76a32f2998473a4fda2e138fe4e4c67c875e8"} Oct 02 16:53:29 crc kubenswrapper[4808]: I1002 16:53:29.515655 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-7p6jd" podUID="d02c0e95-ce3f-42d1-9fb9-9ff040b18447" containerName="registry-server" containerID="cri-o://bd0f92e2bdf532d80ff0baf30c872ef9d4fc1827714aff0feaac0034c67c17da" gracePeriod=2 Oct 02 16:53:29 crc kubenswrapper[4808]: I1002 16:53:29.621197 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2kqnl" Oct 02 16:53:29 crc kubenswrapper[4808]: I1002 16:53:29.621809 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2kqnl" Oct 02 16:53:29 crc kubenswrapper[4808]: I1002 16:53:29.688679 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2kqnl" Oct 02 16:53:30 crc kubenswrapper[4808]: I1002 16:53:30.522388 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7p6jd" Oct 02 16:53:30 crc kubenswrapper[4808]: I1002 16:53:30.524160 4808 generic.go:334] "Generic (PLEG): container finished" podID="d02c0e95-ce3f-42d1-9fb9-9ff040b18447" containerID="bd0f92e2bdf532d80ff0baf30c872ef9d4fc1827714aff0feaac0034c67c17da" exitCode=0 Oct 02 16:53:30 crc kubenswrapper[4808]: I1002 16:53:30.524252 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7p6jd" event={"ID":"d02c0e95-ce3f-42d1-9fb9-9ff040b18447","Type":"ContainerDied","Data":"bd0f92e2bdf532d80ff0baf30c872ef9d4fc1827714aff0feaac0034c67c17da"} Oct 02 16:53:30 crc kubenswrapper[4808]: I1002 16:53:30.524281 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7p6jd" event={"ID":"d02c0e95-ce3f-42d1-9fb9-9ff040b18447","Type":"ContainerDied","Data":"1eeaab89c6fa053096dac6de33882029314e5c618043e9509255bf5f386c47a6"} Oct 02 16:53:30 crc kubenswrapper[4808]: I1002 16:53:30.524293 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7p6jd" Oct 02 16:53:30 crc kubenswrapper[4808]: I1002 16:53:30.524301 4808 scope.go:117] "RemoveContainer" containerID="bd0f92e2bdf532d80ff0baf30c872ef9d4fc1827714aff0feaac0034c67c17da" Oct 02 16:53:30 crc kubenswrapper[4808]: I1002 16:53:30.526093 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zstrp" event={"ID":"1883774e-2d09-400d-95bd-490deded815b","Type":"ContainerStarted","Data":"35d8a29e984ffdf082cfb2c17ec549489939fa8fcb908619ff4afadf45a06de5"} Oct 02 16:53:30 crc kubenswrapper[4808]: I1002 16:53:30.546639 4808 scope.go:117] "RemoveContainer" containerID="bd0f92e2bdf532d80ff0baf30c872ef9d4fc1827714aff0feaac0034c67c17da" Oct 02 16:53:30 crc kubenswrapper[4808]: E1002 16:53:30.547569 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd0f92e2bdf532d80ff0baf30c872ef9d4fc1827714aff0feaac0034c67c17da\": container with ID starting with bd0f92e2bdf532d80ff0baf30c872ef9d4fc1827714aff0feaac0034c67c17da not found: ID does not exist" containerID="bd0f92e2bdf532d80ff0baf30c872ef9d4fc1827714aff0feaac0034c67c17da" Oct 02 16:53:30 crc kubenswrapper[4808]: I1002 16:53:30.547634 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd0f92e2bdf532d80ff0baf30c872ef9d4fc1827714aff0feaac0034c67c17da"} err="failed to get container status \"bd0f92e2bdf532d80ff0baf30c872ef9d4fc1827714aff0feaac0034c67c17da\": rpc error: code = NotFound desc = could not find container \"bd0f92e2bdf532d80ff0baf30c872ef9d4fc1827714aff0feaac0034c67c17da\": container with ID starting with bd0f92e2bdf532d80ff0baf30c872ef9d4fc1827714aff0feaac0034c67c17da not found: ID does not exist" Oct 02 16:53:30 crc kubenswrapper[4808]: I1002 16:53:30.572062 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-zstrp" podStartSLOduration=1.8460694530000001 podStartE2EDuration="2.572034412s" podCreationTimestamp="2025-10-02 16:53:28 +0000 UTC" firstStartedPulling="2025-10-02 16:53:29.286863411 +0000 UTC m=+776.612392411" lastFinishedPulling="2025-10-02 16:53:30.01282837 +0000 UTC m=+777.338357370" observedRunningTime="2025-10-02 16:53:30.562629105 +0000 UTC m=+777.888158125" watchObservedRunningTime="2025-10-02 16:53:30.572034412 +0000 UTC m=+777.897563452" Oct 02 16:53:30 crc kubenswrapper[4808]: I1002 16:53:30.589519 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2kqnl" Oct 02 16:53:30 crc kubenswrapper[4808]: I1002 16:53:30.655287 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zm6h\" (UniqueName: \"kubernetes.io/projected/d02c0e95-ce3f-42d1-9fb9-9ff040b18447-kube-api-access-7zm6h\") pod \"d02c0e95-ce3f-42d1-9fb9-9ff040b18447\" (UID: \"d02c0e95-ce3f-42d1-9fb9-9ff040b18447\") " Oct 02 16:53:30 crc kubenswrapper[4808]: I1002 16:53:30.660401 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d02c0e95-ce3f-42d1-9fb9-9ff040b18447-kube-api-access-7zm6h" (OuterVolumeSpecName: "kube-api-access-7zm6h") pod "d02c0e95-ce3f-42d1-9fb9-9ff040b18447" (UID: "d02c0e95-ce3f-42d1-9fb9-9ff040b18447"). InnerVolumeSpecName "kube-api-access-7zm6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:53:30 crc kubenswrapper[4808]: I1002 16:53:30.758537 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zm6h\" (UniqueName: \"kubernetes.io/projected/d02c0e95-ce3f-42d1-9fb9-9ff040b18447-kube-api-access-7zm6h\") on node \"crc\" DevicePath \"\"" Oct 02 16:53:30 crc kubenswrapper[4808]: I1002 16:53:30.858989 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-7p6jd"] Oct 02 16:53:30 crc kubenswrapper[4808]: I1002 16:53:30.867922 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-7p6jd"] Oct 02 16:53:31 crc kubenswrapper[4808]: I1002 16:53:31.405631 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d02c0e95-ce3f-42d1-9fb9-9ff040b18447" path="/var/lib/kubelet/pods/d02c0e95-ce3f-42d1-9fb9-9ff040b18447/volumes" Oct 02 16:53:33 crc kubenswrapper[4808]: I1002 16:53:33.445005 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2kqnl"] Oct 02 16:53:33 crc kubenswrapper[4808]: I1002 16:53:33.445355 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2kqnl" podUID="d20c4266-42c2-49d2-9fb0-323f5fd208b7" containerName="registry-server" containerID="cri-o://bd334d1eb16c3070282a007c8afd10b65e6b0081d4287ceedbb4af46f72dcaf3" gracePeriod=2 Oct 02 16:53:33 crc kubenswrapper[4808]: I1002 16:53:33.883749 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2kqnl" Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.009227 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d20c4266-42c2-49d2-9fb0-323f5fd208b7-utilities\") pod \"d20c4266-42c2-49d2-9fb0-323f5fd208b7\" (UID: \"d20c4266-42c2-49d2-9fb0-323f5fd208b7\") " Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.009293 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7mxx\" (UniqueName: \"kubernetes.io/projected/d20c4266-42c2-49d2-9fb0-323f5fd208b7-kube-api-access-s7mxx\") pod \"d20c4266-42c2-49d2-9fb0-323f5fd208b7\" (UID: \"d20c4266-42c2-49d2-9fb0-323f5fd208b7\") " Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.009328 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d20c4266-42c2-49d2-9fb0-323f5fd208b7-catalog-content\") pod \"d20c4266-42c2-49d2-9fb0-323f5fd208b7\" (UID: \"d20c4266-42c2-49d2-9fb0-323f5fd208b7\") " Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.010617 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d20c4266-42c2-49d2-9fb0-323f5fd208b7-utilities" (OuterVolumeSpecName: "utilities") pod "d20c4266-42c2-49d2-9fb0-323f5fd208b7" (UID: "d20c4266-42c2-49d2-9fb0-323f5fd208b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.010991 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d20c4266-42c2-49d2-9fb0-323f5fd208b7-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.015688 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d20c4266-42c2-49d2-9fb0-323f5fd208b7-kube-api-access-s7mxx" (OuterVolumeSpecName: "kube-api-access-s7mxx") pod "d20c4266-42c2-49d2-9fb0-323f5fd208b7" (UID: "d20c4266-42c2-49d2-9fb0-323f5fd208b7"). InnerVolumeSpecName "kube-api-access-s7mxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.112169 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7mxx\" (UniqueName: \"kubernetes.io/projected/d20c4266-42c2-49d2-9fb0-323f5fd208b7-kube-api-access-s7mxx\") on node \"crc\" DevicePath \"\"" Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.122132 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d20c4266-42c2-49d2-9fb0-323f5fd208b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d20c4266-42c2-49d2-9fb0-323f5fd208b7" (UID: "d20c4266-42c2-49d2-9fb0-323f5fd208b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.214254 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d20c4266-42c2-49d2-9fb0-323f5fd208b7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.554758 4808 generic.go:334] "Generic (PLEG): container finished" podID="d20c4266-42c2-49d2-9fb0-323f5fd208b7" containerID="bd334d1eb16c3070282a007c8afd10b65e6b0081d4287ceedbb4af46f72dcaf3" exitCode=0 Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.554799 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2kqnl" Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.554806 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2kqnl" event={"ID":"d20c4266-42c2-49d2-9fb0-323f5fd208b7","Type":"ContainerDied","Data":"bd334d1eb16c3070282a007c8afd10b65e6b0081d4287ceedbb4af46f72dcaf3"} Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.554834 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2kqnl" event={"ID":"d20c4266-42c2-49d2-9fb0-323f5fd208b7","Type":"ContainerDied","Data":"1d4d49713581ec695c8efaa1e743f4843e05edad47932676bfe2202c90c16af3"} Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.554852 4808 scope.go:117] "RemoveContainer" containerID="bd334d1eb16c3070282a007c8afd10b65e6b0081d4287ceedbb4af46f72dcaf3" Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.575796 4808 scope.go:117] "RemoveContainer" containerID="fc05e3fce5d6cbd9e6c87112ec4f814d919179e6922f1371e266bad6cc21f707" Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.586468 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2kqnl"] Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.598456 4808 scope.go:117] "RemoveContainer" containerID="2fb106fc6c5d3d1caf8438a9aebebd14ddd789904f3441292b27c8ad53f0815d" Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.601469 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2kqnl"] Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.620791 4808 scope.go:117] "RemoveContainer" containerID="bd334d1eb16c3070282a007c8afd10b65e6b0081d4287ceedbb4af46f72dcaf3" Oct 02 16:53:34 crc kubenswrapper[4808]: E1002 16:53:34.622282 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd334d1eb16c3070282a007c8afd10b65e6b0081d4287ceedbb4af46f72dcaf3\": container with ID starting with bd334d1eb16c3070282a007c8afd10b65e6b0081d4287ceedbb4af46f72dcaf3 not found: ID does not exist" containerID="bd334d1eb16c3070282a007c8afd10b65e6b0081d4287ceedbb4af46f72dcaf3" Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.622347 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd334d1eb16c3070282a007c8afd10b65e6b0081d4287ceedbb4af46f72dcaf3"} err="failed to get container status \"bd334d1eb16c3070282a007c8afd10b65e6b0081d4287ceedbb4af46f72dcaf3\": rpc error: code = NotFound desc = could not find container \"bd334d1eb16c3070282a007c8afd10b65e6b0081d4287ceedbb4af46f72dcaf3\": container with ID starting with bd334d1eb16c3070282a007c8afd10b65e6b0081d4287ceedbb4af46f72dcaf3 not found: ID does not exist" Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.622373 4808 scope.go:117] "RemoveContainer" containerID="fc05e3fce5d6cbd9e6c87112ec4f814d919179e6922f1371e266bad6cc21f707" Oct 02 16:53:34 crc kubenswrapper[4808]: E1002 16:53:34.622783 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc05e3fce5d6cbd9e6c87112ec4f814d919179e6922f1371e266bad6cc21f707\": container with ID starting with fc05e3fce5d6cbd9e6c87112ec4f814d919179e6922f1371e266bad6cc21f707 not found: ID does not exist" containerID="fc05e3fce5d6cbd9e6c87112ec4f814d919179e6922f1371e266bad6cc21f707" Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.622825 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc05e3fce5d6cbd9e6c87112ec4f814d919179e6922f1371e266bad6cc21f707"} err="failed to get container status \"fc05e3fce5d6cbd9e6c87112ec4f814d919179e6922f1371e266bad6cc21f707\": rpc error: code = NotFound desc = could not find container \"fc05e3fce5d6cbd9e6c87112ec4f814d919179e6922f1371e266bad6cc21f707\": container with ID starting with fc05e3fce5d6cbd9e6c87112ec4f814d919179e6922f1371e266bad6cc21f707 not found: ID does not exist" Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.622855 4808 scope.go:117] "RemoveContainer" containerID="2fb106fc6c5d3d1caf8438a9aebebd14ddd789904f3441292b27c8ad53f0815d" Oct 02 16:53:34 crc kubenswrapper[4808]: E1002 16:53:34.623057 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fb106fc6c5d3d1caf8438a9aebebd14ddd789904f3441292b27c8ad53f0815d\": container with ID starting with 2fb106fc6c5d3d1caf8438a9aebebd14ddd789904f3441292b27c8ad53f0815d not found: ID does not exist" containerID="2fb106fc6c5d3d1caf8438a9aebebd14ddd789904f3441292b27c8ad53f0815d" Oct 02 16:53:34 crc kubenswrapper[4808]: I1002 16:53:34.623084 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fb106fc6c5d3d1caf8438a9aebebd14ddd789904f3441292b27c8ad53f0815d"} err="failed to get container status \"2fb106fc6c5d3d1caf8438a9aebebd14ddd789904f3441292b27c8ad53f0815d\": rpc error: code = NotFound desc = could not find container \"2fb106fc6c5d3d1caf8438a9aebebd14ddd789904f3441292b27c8ad53f0815d\": container with ID starting with 2fb106fc6c5d3d1caf8438a9aebebd14ddd789904f3441292b27c8ad53f0815d not found: ID does not exist" Oct 02 16:53:35 crc kubenswrapper[4808]: I1002 16:53:35.409716 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d20c4266-42c2-49d2-9fb0-323f5fd208b7" path="/var/lib/kubelet/pods/d20c4266-42c2-49d2-9fb0-323f5fd208b7/volumes" Oct 02 16:53:38 crc kubenswrapper[4808]: I1002 16:53:38.785077 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-zstrp" Oct 02 16:53:38 crc kubenswrapper[4808]: I1002 16:53:38.785614 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-zstrp" Oct 02 16:53:38 crc kubenswrapper[4808]: I1002 16:53:38.826340 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-zstrp" Oct 02 16:53:39 crc kubenswrapper[4808]: I1002 16:53:39.635124 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-zstrp" Oct 02 16:53:45 crc kubenswrapper[4808]: I1002 16:53:45.701701 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf"] Oct 02 16:53:45 crc kubenswrapper[4808]: E1002 16:53:45.704193 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d20c4266-42c2-49d2-9fb0-323f5fd208b7" containerName="extract-content" Oct 02 16:53:45 crc kubenswrapper[4808]: I1002 16:53:45.704249 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d20c4266-42c2-49d2-9fb0-323f5fd208b7" containerName="extract-content" Oct 02 16:53:45 crc kubenswrapper[4808]: E1002 16:53:45.704264 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d20c4266-42c2-49d2-9fb0-323f5fd208b7" containerName="extract-utilities" Oct 02 16:53:45 crc kubenswrapper[4808]: I1002 16:53:45.704273 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d20c4266-42c2-49d2-9fb0-323f5fd208b7" containerName="extract-utilities" Oct 02 16:53:45 crc kubenswrapper[4808]: E1002 16:53:45.704291 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d20c4266-42c2-49d2-9fb0-323f5fd208b7" containerName="registry-server" Oct 02 16:53:45 crc kubenswrapper[4808]: I1002 16:53:45.704299 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d20c4266-42c2-49d2-9fb0-323f5fd208b7" containerName="registry-server" Oct 02 16:53:45 crc kubenswrapper[4808]: E1002 16:53:45.704318 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d02c0e95-ce3f-42d1-9fb9-9ff040b18447" containerName="registry-server" Oct 02 16:53:45 crc kubenswrapper[4808]: I1002 16:53:45.704329 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d02c0e95-ce3f-42d1-9fb9-9ff040b18447" containerName="registry-server" Oct 02 16:53:45 crc kubenswrapper[4808]: I1002 16:53:45.704482 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="d20c4266-42c2-49d2-9fb0-323f5fd208b7" containerName="registry-server" Oct 02 16:53:45 crc kubenswrapper[4808]: I1002 16:53:45.704500 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="d02c0e95-ce3f-42d1-9fb9-9ff040b18447" containerName="registry-server" Oct 02 16:53:45 crc kubenswrapper[4808]: I1002 16:53:45.705650 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf" Oct 02 16:53:45 crc kubenswrapper[4808]: I1002 16:53:45.709022 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-485gx" Oct 02 16:53:45 crc kubenswrapper[4808]: I1002 16:53:45.718860 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf"] Oct 02 16:53:45 crc kubenswrapper[4808]: I1002 16:53:45.887119 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcr6z\" (UniqueName: \"kubernetes.io/projected/8771d707-d0d8-4b3b-8ca5-d68949b29b64-kube-api-access-xcr6z\") pod \"1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf\" (UID: \"8771d707-d0d8-4b3b-8ca5-d68949b29b64\") " pod="openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf" Oct 02 16:53:45 crc kubenswrapper[4808]: I1002 16:53:45.887168 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8771d707-d0d8-4b3b-8ca5-d68949b29b64-util\") pod \"1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf\" (UID: \"8771d707-d0d8-4b3b-8ca5-d68949b29b64\") " pod="openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf" Oct 02 16:53:45 crc kubenswrapper[4808]: I1002 16:53:45.887195 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8771d707-d0d8-4b3b-8ca5-d68949b29b64-bundle\") pod \"1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf\" (UID: \"8771d707-d0d8-4b3b-8ca5-d68949b29b64\") " pod="openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf" Oct 02 16:53:45 crc kubenswrapper[4808]: I1002 16:53:45.988053 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcr6z\" (UniqueName: \"kubernetes.io/projected/8771d707-d0d8-4b3b-8ca5-d68949b29b64-kube-api-access-xcr6z\") pod \"1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf\" (UID: \"8771d707-d0d8-4b3b-8ca5-d68949b29b64\") " pod="openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf" Oct 02 16:53:45 crc kubenswrapper[4808]: I1002 16:53:45.988113 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8771d707-d0d8-4b3b-8ca5-d68949b29b64-util\") pod \"1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf\" (UID: \"8771d707-d0d8-4b3b-8ca5-d68949b29b64\") " pod="openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf" Oct 02 16:53:45 crc kubenswrapper[4808]: I1002 16:53:45.988146 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8771d707-d0d8-4b3b-8ca5-d68949b29b64-bundle\") pod \"1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf\" (UID: \"8771d707-d0d8-4b3b-8ca5-d68949b29b64\") " pod="openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf" Oct 02 16:53:45 crc kubenswrapper[4808]: I1002 16:53:45.988610 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8771d707-d0d8-4b3b-8ca5-d68949b29b64-util\") pod \"1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf\" (UID: \"8771d707-d0d8-4b3b-8ca5-d68949b29b64\") " pod="openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf" Oct 02 16:53:45 crc kubenswrapper[4808]: I1002 16:53:45.988627 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8771d707-d0d8-4b3b-8ca5-d68949b29b64-bundle\") pod \"1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf\" (UID: \"8771d707-d0d8-4b3b-8ca5-d68949b29b64\") " pod="openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf" Oct 02 16:53:46 crc kubenswrapper[4808]: I1002 16:53:46.013646 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcr6z\" (UniqueName: \"kubernetes.io/projected/8771d707-d0d8-4b3b-8ca5-d68949b29b64-kube-api-access-xcr6z\") pod \"1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf\" (UID: \"8771d707-d0d8-4b3b-8ca5-d68949b29b64\") " pod="openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf" Oct 02 16:53:46 crc kubenswrapper[4808]: I1002 16:53:46.029354 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf" Oct 02 16:53:46 crc kubenswrapper[4808]: I1002 16:53:46.425841 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf"] Oct 02 16:53:46 crc kubenswrapper[4808]: I1002 16:53:46.651100 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf" event={"ID":"8771d707-d0d8-4b3b-8ca5-d68949b29b64","Type":"ContainerStarted","Data":"115151a81262d711959aa6e63f0951f807617bd3353aca9de8a3b647fc568602"} Oct 02 16:53:46 crc kubenswrapper[4808]: I1002 16:53:46.651595 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf" event={"ID":"8771d707-d0d8-4b3b-8ca5-d68949b29b64","Type":"ContainerStarted","Data":"2631940f0343f4c00355538edf0cc219d22d893b5c17595e794c676e60e3c330"} Oct 02 16:53:47 crc kubenswrapper[4808]: I1002 16:53:47.661048 4808 generic.go:334] "Generic (PLEG): container finished" podID="8771d707-d0d8-4b3b-8ca5-d68949b29b64" containerID="115151a81262d711959aa6e63f0951f807617bd3353aca9de8a3b647fc568602" exitCode=0 Oct 02 16:53:47 crc kubenswrapper[4808]: I1002 16:53:47.661098 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf" event={"ID":"8771d707-d0d8-4b3b-8ca5-d68949b29b64","Type":"ContainerDied","Data":"115151a81262d711959aa6e63f0951f807617bd3353aca9de8a3b647fc568602"} Oct 02 16:53:48 crc kubenswrapper[4808]: I1002 16:53:48.668433 4808 generic.go:334] "Generic (PLEG): container finished" podID="8771d707-d0d8-4b3b-8ca5-d68949b29b64" containerID="5cec2fcef222c25724d19923f9111fba16f2c6aad8805d739106f7816512f2c2" exitCode=0 Oct 02 16:53:48 crc kubenswrapper[4808]: I1002 16:53:48.668667 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf" event={"ID":"8771d707-d0d8-4b3b-8ca5-d68949b29b64","Type":"ContainerDied","Data":"5cec2fcef222c25724d19923f9111fba16f2c6aad8805d739106f7816512f2c2"} Oct 02 16:53:48 crc kubenswrapper[4808]: I1002 16:53:48.739888 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 16:53:48 crc kubenswrapper[4808]: I1002 16:53:48.739942 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 16:53:49 crc kubenswrapper[4808]: I1002 16:53:49.677661 4808 generic.go:334] "Generic (PLEG): container finished" podID="8771d707-d0d8-4b3b-8ca5-d68949b29b64" containerID="9fae6c1c2ff95b5fe6121c9e6aae97380f4a249337e0f2de3dfb6f660c5c4fc6" exitCode=0 Oct 02 16:53:49 crc kubenswrapper[4808]: I1002 16:53:49.677777 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf" event={"ID":"8771d707-d0d8-4b3b-8ca5-d68949b29b64","Type":"ContainerDied","Data":"9fae6c1c2ff95b5fe6121c9e6aae97380f4a249337e0f2de3dfb6f660c5c4fc6"} Oct 02 16:53:50 crc kubenswrapper[4808]: I1002 16:53:50.990712 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf" Oct 02 16:53:51 crc kubenswrapper[4808]: I1002 16:53:51.158028 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8771d707-d0d8-4b3b-8ca5-d68949b29b64-util\") pod \"8771d707-d0d8-4b3b-8ca5-d68949b29b64\" (UID: \"8771d707-d0d8-4b3b-8ca5-d68949b29b64\") " Oct 02 16:53:51 crc kubenswrapper[4808]: I1002 16:53:51.158097 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcr6z\" (UniqueName: \"kubernetes.io/projected/8771d707-d0d8-4b3b-8ca5-d68949b29b64-kube-api-access-xcr6z\") pod \"8771d707-d0d8-4b3b-8ca5-d68949b29b64\" (UID: \"8771d707-d0d8-4b3b-8ca5-d68949b29b64\") " Oct 02 16:53:51 crc kubenswrapper[4808]: I1002 16:53:51.158251 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8771d707-d0d8-4b3b-8ca5-d68949b29b64-bundle\") pod \"8771d707-d0d8-4b3b-8ca5-d68949b29b64\" (UID: \"8771d707-d0d8-4b3b-8ca5-d68949b29b64\") " Oct 02 16:53:51 crc kubenswrapper[4808]: I1002 16:53:51.159124 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8771d707-d0d8-4b3b-8ca5-d68949b29b64-bundle" (OuterVolumeSpecName: "bundle") pod "8771d707-d0d8-4b3b-8ca5-d68949b29b64" (UID: "8771d707-d0d8-4b3b-8ca5-d68949b29b64"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:53:51 crc kubenswrapper[4808]: I1002 16:53:51.169755 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8771d707-d0d8-4b3b-8ca5-d68949b29b64-kube-api-access-xcr6z" (OuterVolumeSpecName: "kube-api-access-xcr6z") pod "8771d707-d0d8-4b3b-8ca5-d68949b29b64" (UID: "8771d707-d0d8-4b3b-8ca5-d68949b29b64"). InnerVolumeSpecName "kube-api-access-xcr6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:53:51 crc kubenswrapper[4808]: I1002 16:53:51.171793 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8771d707-d0d8-4b3b-8ca5-d68949b29b64-util" (OuterVolumeSpecName: "util") pod "8771d707-d0d8-4b3b-8ca5-d68949b29b64" (UID: "8771d707-d0d8-4b3b-8ca5-d68949b29b64"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:53:51 crc kubenswrapper[4808]: I1002 16:53:51.259935 4808 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8771d707-d0d8-4b3b-8ca5-d68949b29b64-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:53:51 crc kubenswrapper[4808]: I1002 16:53:51.259971 4808 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8771d707-d0d8-4b3b-8ca5-d68949b29b64-util\") on node \"crc\" DevicePath \"\"" Oct 02 16:53:51 crc kubenswrapper[4808]: I1002 16:53:51.259986 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcr6z\" (UniqueName: \"kubernetes.io/projected/8771d707-d0d8-4b3b-8ca5-d68949b29b64-kube-api-access-xcr6z\") on node \"crc\" DevicePath \"\"" Oct 02 16:53:51 crc kubenswrapper[4808]: I1002 16:53:51.693694 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf" event={"ID":"8771d707-d0d8-4b3b-8ca5-d68949b29b64","Type":"ContainerDied","Data":"2631940f0343f4c00355538edf0cc219d22d893b5c17595e794c676e60e3c330"} Oct 02 16:53:51 crc kubenswrapper[4808]: I1002 16:53:51.693952 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2631940f0343f4c00355538edf0cc219d22d893b5c17595e794c676e60e3c330" Oct 02 16:53:51 crc kubenswrapper[4808]: I1002 16:53:51.693776 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf" Oct 02 16:53:56 crc kubenswrapper[4808]: I1002 16:53:56.179570 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-755876968d-m4vm4"] Oct 02 16:53:56 crc kubenswrapper[4808]: E1002 16:53:56.181604 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8771d707-d0d8-4b3b-8ca5-d68949b29b64" containerName="util" Oct 02 16:53:56 crc kubenswrapper[4808]: I1002 16:53:56.181706 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8771d707-d0d8-4b3b-8ca5-d68949b29b64" containerName="util" Oct 02 16:53:56 crc kubenswrapper[4808]: E1002 16:53:56.181800 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8771d707-d0d8-4b3b-8ca5-d68949b29b64" containerName="extract" Oct 02 16:53:56 crc kubenswrapper[4808]: I1002 16:53:56.181877 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8771d707-d0d8-4b3b-8ca5-d68949b29b64" containerName="extract" Oct 02 16:53:56 crc kubenswrapper[4808]: E1002 16:53:56.181956 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8771d707-d0d8-4b3b-8ca5-d68949b29b64" containerName="pull" Oct 02 16:53:56 crc kubenswrapper[4808]: I1002 16:53:56.182032 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8771d707-d0d8-4b3b-8ca5-d68949b29b64" containerName="pull" Oct 02 16:53:56 crc kubenswrapper[4808]: I1002 16:53:56.182484 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="8771d707-d0d8-4b3b-8ca5-d68949b29b64" containerName="extract" Oct 02 16:53:56 crc kubenswrapper[4808]: I1002 16:53:56.183361 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-755876968d-m4vm4" Oct 02 16:53:56 crc kubenswrapper[4808]: I1002 16:53:56.186809 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-5d2j4" Oct 02 16:53:56 crc kubenswrapper[4808]: I1002 16:53:56.282444 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-755876968d-m4vm4"] Oct 02 16:53:56 crc kubenswrapper[4808]: I1002 16:53:56.328955 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgrjn\" (UniqueName: \"kubernetes.io/projected/b0800622-881c-4e2a-9cb1-0447445eddc1-kube-api-access-qgrjn\") pod \"openstack-operator-controller-operator-755876968d-m4vm4\" (UID: \"b0800622-881c-4e2a-9cb1-0447445eddc1\") " pod="openstack-operators/openstack-operator-controller-operator-755876968d-m4vm4" Oct 02 16:53:56 crc kubenswrapper[4808]: I1002 16:53:56.430962 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgrjn\" (UniqueName: \"kubernetes.io/projected/b0800622-881c-4e2a-9cb1-0447445eddc1-kube-api-access-qgrjn\") pod \"openstack-operator-controller-operator-755876968d-m4vm4\" (UID: \"b0800622-881c-4e2a-9cb1-0447445eddc1\") " pod="openstack-operators/openstack-operator-controller-operator-755876968d-m4vm4" Oct 02 16:53:56 crc kubenswrapper[4808]: I1002 16:53:56.453076 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgrjn\" (UniqueName: \"kubernetes.io/projected/b0800622-881c-4e2a-9cb1-0447445eddc1-kube-api-access-qgrjn\") pod \"openstack-operator-controller-operator-755876968d-m4vm4\" (UID: \"b0800622-881c-4e2a-9cb1-0447445eddc1\") " pod="openstack-operators/openstack-operator-controller-operator-755876968d-m4vm4" Oct 02 16:53:56 crc kubenswrapper[4808]: I1002 16:53:56.503534 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-755876968d-m4vm4" Oct 02 16:53:56 crc kubenswrapper[4808]: I1002 16:53:56.920035 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-755876968d-m4vm4"] Oct 02 16:53:57 crc kubenswrapper[4808]: I1002 16:53:57.728980 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-755876968d-m4vm4" event={"ID":"b0800622-881c-4e2a-9cb1-0447445eddc1","Type":"ContainerStarted","Data":"db48c80ebc6cff0b8018dfe44e7a7c7cff074925ba92e98bb9666367dc964ec2"} Oct 02 16:54:01 crc kubenswrapper[4808]: I1002 16:54:01.757530 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-755876968d-m4vm4" event={"ID":"b0800622-881c-4e2a-9cb1-0447445eddc1","Type":"ContainerStarted","Data":"64d5cab6e2beef53be93537e13b1a9d907696a215677ee61be3c82aa4105bcb6"} Oct 02 16:54:04 crc kubenswrapper[4808]: I1002 16:54:04.777963 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-755876968d-m4vm4" event={"ID":"b0800622-881c-4e2a-9cb1-0447445eddc1","Type":"ContainerStarted","Data":"a1970b896a1bacb481d24cbcbb998e4d9ecd0c65d8e6ad8f4c9108e4b70cc112"} Oct 02 16:54:04 crc kubenswrapper[4808]: I1002 16:54:04.778578 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-755876968d-m4vm4" Oct 02 16:54:04 crc kubenswrapper[4808]: I1002 16:54:04.814093 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-755876968d-m4vm4" podStartSLOduration=1.824715924 podStartE2EDuration="8.814074683s" podCreationTimestamp="2025-10-02 16:53:56 +0000 UTC" firstStartedPulling="2025-10-02 16:53:56.92546871 +0000 UTC m=+804.250997710" lastFinishedPulling="2025-10-02 16:54:03.914827459 +0000 UTC m=+811.240356469" observedRunningTime="2025-10-02 16:54:04.81172602 +0000 UTC m=+812.137255030" watchObservedRunningTime="2025-10-02 16:54:04.814074683 +0000 UTC m=+812.139603693" Oct 02 16:54:06 crc kubenswrapper[4808]: I1002 16:54:06.507215 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-755876968d-m4vm4" Oct 02 16:54:15 crc kubenswrapper[4808]: I1002 16:54:15.034744 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h2tkr"] Oct 02 16:54:15 crc kubenswrapper[4808]: I1002 16:54:15.037153 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h2tkr" Oct 02 16:54:15 crc kubenswrapper[4808]: I1002 16:54:15.047381 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h2tkr"] Oct 02 16:54:15 crc kubenswrapper[4808]: I1002 16:54:15.093085 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fba922bf-eff6-4b8a-92bf-8a9bdce0c83f-utilities\") pod \"redhat-marketplace-h2tkr\" (UID: \"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f\") " pod="openshift-marketplace/redhat-marketplace-h2tkr" Oct 02 16:54:15 crc kubenswrapper[4808]: I1002 16:54:15.093258 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fba922bf-eff6-4b8a-92bf-8a9bdce0c83f-catalog-content\") pod \"redhat-marketplace-h2tkr\" (UID: \"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f\") " pod="openshift-marketplace/redhat-marketplace-h2tkr" Oct 02 16:54:15 crc kubenswrapper[4808]: I1002 16:54:15.093383 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqh56\" (UniqueName: \"kubernetes.io/projected/fba922bf-eff6-4b8a-92bf-8a9bdce0c83f-kube-api-access-rqh56\") pod \"redhat-marketplace-h2tkr\" (UID: \"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f\") " pod="openshift-marketplace/redhat-marketplace-h2tkr" Oct 02 16:54:15 crc kubenswrapper[4808]: I1002 16:54:15.194503 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fba922bf-eff6-4b8a-92bf-8a9bdce0c83f-utilities\") pod \"redhat-marketplace-h2tkr\" (UID: \"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f\") " pod="openshift-marketplace/redhat-marketplace-h2tkr" Oct 02 16:54:15 crc kubenswrapper[4808]: I1002 16:54:15.194637 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fba922bf-eff6-4b8a-92bf-8a9bdce0c83f-catalog-content\") pod \"redhat-marketplace-h2tkr\" (UID: \"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f\") " pod="openshift-marketplace/redhat-marketplace-h2tkr" Oct 02 16:54:15 crc kubenswrapper[4808]: I1002 16:54:15.194696 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqh56\" (UniqueName: \"kubernetes.io/projected/fba922bf-eff6-4b8a-92bf-8a9bdce0c83f-kube-api-access-rqh56\") pod \"redhat-marketplace-h2tkr\" (UID: \"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f\") " pod="openshift-marketplace/redhat-marketplace-h2tkr" Oct 02 16:54:15 crc kubenswrapper[4808]: I1002 16:54:15.196212 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fba922bf-eff6-4b8a-92bf-8a9bdce0c83f-utilities\") pod \"redhat-marketplace-h2tkr\" (UID: \"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f\") " pod="openshift-marketplace/redhat-marketplace-h2tkr" Oct 02 16:54:15 crc kubenswrapper[4808]: I1002 16:54:15.196629 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fba922bf-eff6-4b8a-92bf-8a9bdce0c83f-catalog-content\") pod \"redhat-marketplace-h2tkr\" (UID: \"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f\") " pod="openshift-marketplace/redhat-marketplace-h2tkr" Oct 02 16:54:15 crc kubenswrapper[4808]: I1002 16:54:15.223876 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqh56\" (UniqueName: \"kubernetes.io/projected/fba922bf-eff6-4b8a-92bf-8a9bdce0c83f-kube-api-access-rqh56\") pod \"redhat-marketplace-h2tkr\" (UID: \"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f\") " pod="openshift-marketplace/redhat-marketplace-h2tkr" Oct 02 16:54:15 crc kubenswrapper[4808]: I1002 16:54:15.352675 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h2tkr" Oct 02 16:54:15 crc kubenswrapper[4808]: I1002 16:54:15.799909 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h2tkr"] Oct 02 16:54:15 crc kubenswrapper[4808]: I1002 16:54:15.855170 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2tkr" event={"ID":"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f","Type":"ContainerStarted","Data":"b30e6e9856f6e2e686fc0acf2086e9db449501cab42394a0af9fd76cfc63f22f"} Oct 02 16:54:16 crc kubenswrapper[4808]: I1002 16:54:16.864798 4808 generic.go:334] "Generic (PLEG): container finished" podID="fba922bf-eff6-4b8a-92bf-8a9bdce0c83f" containerID="b5b6497c917a202544feed1fd2ea96bc26e272644b86a4b79cdf77ffadc0b377" exitCode=0 Oct 02 16:54:16 crc kubenswrapper[4808]: I1002 16:54:16.864857 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2tkr" event={"ID":"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f","Type":"ContainerDied","Data":"b5b6497c917a202544feed1fd2ea96bc26e272644b86a4b79cdf77ffadc0b377"} Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.028498 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jz72l"] Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.030790 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jz72l" Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.037850 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8w6b\" (UniqueName: \"kubernetes.io/projected/21011486-b620-4b60-9b06-a7e3e003a37d-kube-api-access-b8w6b\") pod \"community-operators-jz72l\" (UID: \"21011486-b620-4b60-9b06-a7e3e003a37d\") " pod="openshift-marketplace/community-operators-jz72l" Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.038008 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21011486-b620-4b60-9b06-a7e3e003a37d-catalog-content\") pod \"community-operators-jz72l\" (UID: \"21011486-b620-4b60-9b06-a7e3e003a37d\") " pod="openshift-marketplace/community-operators-jz72l" Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.038084 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21011486-b620-4b60-9b06-a7e3e003a37d-utilities\") pod \"community-operators-jz72l\" (UID: \"21011486-b620-4b60-9b06-a7e3e003a37d\") " pod="openshift-marketplace/community-operators-jz72l" Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.038088 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jz72l"] Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.139599 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21011486-b620-4b60-9b06-a7e3e003a37d-catalog-content\") pod \"community-operators-jz72l\" (UID: \"21011486-b620-4b60-9b06-a7e3e003a37d\") " pod="openshift-marketplace/community-operators-jz72l" Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.139656 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21011486-b620-4b60-9b06-a7e3e003a37d-utilities\") pod \"community-operators-jz72l\" (UID: \"21011486-b620-4b60-9b06-a7e3e003a37d\") " pod="openshift-marketplace/community-operators-jz72l" Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.140105 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8w6b\" (UniqueName: \"kubernetes.io/projected/21011486-b620-4b60-9b06-a7e3e003a37d-kube-api-access-b8w6b\") pod \"community-operators-jz72l\" (UID: \"21011486-b620-4b60-9b06-a7e3e003a37d\") " pod="openshift-marketplace/community-operators-jz72l" Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.140217 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21011486-b620-4b60-9b06-a7e3e003a37d-utilities\") pod \"community-operators-jz72l\" (UID: \"21011486-b620-4b60-9b06-a7e3e003a37d\") " pod="openshift-marketplace/community-operators-jz72l" Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.140633 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21011486-b620-4b60-9b06-a7e3e003a37d-catalog-content\") pod \"community-operators-jz72l\" (UID: \"21011486-b620-4b60-9b06-a7e3e003a37d\") " pod="openshift-marketplace/community-operators-jz72l" Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.172834 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8w6b\" (UniqueName: \"kubernetes.io/projected/21011486-b620-4b60-9b06-a7e3e003a37d-kube-api-access-b8w6b\") pod \"community-operators-jz72l\" (UID: \"21011486-b620-4b60-9b06-a7e3e003a37d\") " pod="openshift-marketplace/community-operators-jz72l" Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.363352 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jz72l" Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.624962 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jz72l"] Oct 02 16:54:18 crc kubenswrapper[4808]: W1002 16:54:18.641642 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21011486_b620_4b60_9b06_a7e3e003a37d.slice/crio-1cc0c05dd409ef286f4d1f67c9f7404b9cd7fc13ee5caca6130138c1ff90623d WatchSource:0}: Error finding container 1cc0c05dd409ef286f4d1f67c9f7404b9cd7fc13ee5caca6130138c1ff90623d: Status 404 returned error can't find the container with id 1cc0c05dd409ef286f4d1f67c9f7404b9cd7fc13ee5caca6130138c1ff90623d Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.739265 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.739322 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.883529 4808 generic.go:334] "Generic (PLEG): container finished" podID="fba922bf-eff6-4b8a-92bf-8a9bdce0c83f" containerID="b0d1b1f129a158100c58b420883b6d114d282f2f1ef8acb200f051d4d4f86022" exitCode=0 Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.883592 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2tkr" event={"ID":"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f","Type":"ContainerDied","Data":"b0d1b1f129a158100c58b420883b6d114d282f2f1ef8acb200f051d4d4f86022"} Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.884909 4808 generic.go:334] "Generic (PLEG): container finished" podID="21011486-b620-4b60-9b06-a7e3e003a37d" containerID="05fe5a1df9a5d8e80cd7c17cdda684ac4ffba3071c4fe6e86777f643b8773dc1" exitCode=0 Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.884930 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jz72l" event={"ID":"21011486-b620-4b60-9b06-a7e3e003a37d","Type":"ContainerDied","Data":"05fe5a1df9a5d8e80cd7c17cdda684ac4ffba3071c4fe6e86777f643b8773dc1"} Oct 02 16:54:18 crc kubenswrapper[4808]: I1002 16:54:18.884943 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jz72l" event={"ID":"21011486-b620-4b60-9b06-a7e3e003a37d","Type":"ContainerStarted","Data":"1cc0c05dd409ef286f4d1f67c9f7404b9cd7fc13ee5caca6130138c1ff90623d"} Oct 02 16:54:20 crc kubenswrapper[4808]: I1002 16:54:20.903565 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2tkr" event={"ID":"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f","Type":"ContainerStarted","Data":"e5cfb4006a1eef69a5dd5a89cac0731ae5a7342cf2558cb5c049a15ab849b1cc"} Oct 02 16:54:20 crc kubenswrapper[4808]: I1002 16:54:20.905608 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jz72l" event={"ID":"21011486-b620-4b60-9b06-a7e3e003a37d","Type":"ContainerStarted","Data":"fdd4714c2c77183454a28ee4a6fc8f9e20eb123c77353659006b7304b6568fb6"} Oct 02 16:54:20 crc kubenswrapper[4808]: I1002 16:54:20.925069 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h2tkr" podStartSLOduration=3.064241678 podStartE2EDuration="5.925034966s" podCreationTimestamp="2025-10-02 16:54:15 +0000 UTC" firstStartedPulling="2025-10-02 16:54:16.867118288 +0000 UTC m=+824.192647298" lastFinishedPulling="2025-10-02 16:54:19.727911576 +0000 UTC m=+827.053440586" observedRunningTime="2025-10-02 16:54:20.924481981 +0000 UTC m=+828.250011021" watchObservedRunningTime="2025-10-02 16:54:20.925034966 +0000 UTC m=+828.250564016" Oct 02 16:54:21 crc kubenswrapper[4808]: I1002 16:54:21.918340 4808 generic.go:334] "Generic (PLEG): container finished" podID="21011486-b620-4b60-9b06-a7e3e003a37d" containerID="fdd4714c2c77183454a28ee4a6fc8f9e20eb123c77353659006b7304b6568fb6" exitCode=0 Oct 02 16:54:21 crc kubenswrapper[4808]: I1002 16:54:21.918523 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jz72l" event={"ID":"21011486-b620-4b60-9b06-a7e3e003a37d","Type":"ContainerDied","Data":"fdd4714c2c77183454a28ee4a6fc8f9e20eb123c77353659006b7304b6568fb6"} Oct 02 16:54:22 crc kubenswrapper[4808]: I1002 16:54:22.929491 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jz72l" event={"ID":"21011486-b620-4b60-9b06-a7e3e003a37d","Type":"ContainerStarted","Data":"7246b42ea3a14518a2ebcbe0c490937720a4337719139a619153aba07c2d2169"} Oct 02 16:54:22 crc kubenswrapper[4808]: I1002 16:54:22.949931 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jz72l" podStartSLOduration=3.250699445 podStartE2EDuration="5.94990852s" podCreationTimestamp="2025-10-02 16:54:17 +0000 UTC" firstStartedPulling="2025-10-02 16:54:19.897293101 +0000 UTC m=+827.222822141" lastFinishedPulling="2025-10-02 16:54:22.596502186 +0000 UTC m=+829.922031216" observedRunningTime="2025-10-02 16:54:22.947036722 +0000 UTC m=+830.272565732" watchObservedRunningTime="2025-10-02 16:54:22.94990852 +0000 UTC m=+830.275437540" Oct 02 16:54:25 crc kubenswrapper[4808]: I1002 16:54:25.353789 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h2tkr" Oct 02 16:54:25 crc kubenswrapper[4808]: I1002 16:54:25.354315 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h2tkr" Oct 02 16:54:25 crc kubenswrapper[4808]: I1002 16:54:25.406959 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h2tkr" Oct 02 16:54:26 crc kubenswrapper[4808]: I1002 16:54:26.033570 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h2tkr" Oct 02 16:54:26 crc kubenswrapper[4808]: I1002 16:54:26.612641 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h2tkr"] Oct 02 16:54:27 crc kubenswrapper[4808]: I1002 16:54:27.995012 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h2tkr" podUID="fba922bf-eff6-4b8a-92bf-8a9bdce0c83f" containerName="registry-server" containerID="cri-o://e5cfb4006a1eef69a5dd5a89cac0731ae5a7342cf2558cb5c049a15ab849b1cc" gracePeriod=2 Oct 02 16:54:28 crc kubenswrapper[4808]: I1002 16:54:28.364257 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jz72l" Oct 02 16:54:28 crc kubenswrapper[4808]: I1002 16:54:28.364349 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jz72l" Oct 02 16:54:28 crc kubenswrapper[4808]: I1002 16:54:28.426606 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jz72l" Oct 02 16:54:29 crc kubenswrapper[4808]: I1002 16:54:29.001813 4808 generic.go:334] "Generic (PLEG): container finished" podID="fba922bf-eff6-4b8a-92bf-8a9bdce0c83f" containerID="e5cfb4006a1eef69a5dd5a89cac0731ae5a7342cf2558cb5c049a15ab849b1cc" exitCode=0 Oct 02 16:54:29 crc kubenswrapper[4808]: I1002 16:54:29.001835 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2tkr" event={"ID":"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f","Type":"ContainerDied","Data":"e5cfb4006a1eef69a5dd5a89cac0731ae5a7342cf2558cb5c049a15ab849b1cc"} Oct 02 16:54:29 crc kubenswrapper[4808]: I1002 16:54:29.067637 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jz72l" Oct 02 16:54:30 crc kubenswrapper[4808]: I1002 16:54:30.211524 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jz72l"] Oct 02 16:54:30 crc kubenswrapper[4808]: I1002 16:54:30.341317 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h2tkr" Oct 02 16:54:30 crc kubenswrapper[4808]: I1002 16:54:30.522823 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fba922bf-eff6-4b8a-92bf-8a9bdce0c83f-utilities\") pod \"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f\" (UID: \"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f\") " Oct 02 16:54:30 crc kubenswrapper[4808]: I1002 16:54:30.522883 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fba922bf-eff6-4b8a-92bf-8a9bdce0c83f-catalog-content\") pod \"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f\" (UID: \"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f\") " Oct 02 16:54:30 crc kubenswrapper[4808]: I1002 16:54:30.522911 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqh56\" (UniqueName: \"kubernetes.io/projected/fba922bf-eff6-4b8a-92bf-8a9bdce0c83f-kube-api-access-rqh56\") pod \"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f\" (UID: \"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f\") " Oct 02 16:54:30 crc kubenswrapper[4808]: I1002 16:54:30.524521 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fba922bf-eff6-4b8a-92bf-8a9bdce0c83f-utilities" (OuterVolumeSpecName: "utilities") pod "fba922bf-eff6-4b8a-92bf-8a9bdce0c83f" (UID: "fba922bf-eff6-4b8a-92bf-8a9bdce0c83f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:54:30 crc kubenswrapper[4808]: I1002 16:54:30.532386 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fba922bf-eff6-4b8a-92bf-8a9bdce0c83f-kube-api-access-rqh56" (OuterVolumeSpecName: "kube-api-access-rqh56") pod "fba922bf-eff6-4b8a-92bf-8a9bdce0c83f" (UID: "fba922bf-eff6-4b8a-92bf-8a9bdce0c83f"). InnerVolumeSpecName "kube-api-access-rqh56". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:54:30 crc kubenswrapper[4808]: I1002 16:54:30.539592 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fba922bf-eff6-4b8a-92bf-8a9bdce0c83f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fba922bf-eff6-4b8a-92bf-8a9bdce0c83f" (UID: "fba922bf-eff6-4b8a-92bf-8a9bdce0c83f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:54:30 crc kubenswrapper[4808]: I1002 16:54:30.623873 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fba922bf-eff6-4b8a-92bf-8a9bdce0c83f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 16:54:30 crc kubenswrapper[4808]: I1002 16:54:30.623909 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqh56\" (UniqueName: \"kubernetes.io/projected/fba922bf-eff6-4b8a-92bf-8a9bdce0c83f-kube-api-access-rqh56\") on node \"crc\" DevicePath \"\"" Oct 02 16:54:30 crc kubenswrapper[4808]: I1002 16:54:30.623922 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fba922bf-eff6-4b8a-92bf-8a9bdce0c83f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 16:54:31 crc kubenswrapper[4808]: I1002 16:54:31.017591 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2tkr" event={"ID":"fba922bf-eff6-4b8a-92bf-8a9bdce0c83f","Type":"ContainerDied","Data":"b30e6e9856f6e2e686fc0acf2086e9db449501cab42394a0af9fd76cfc63f22f"} Oct 02 16:54:31 crc kubenswrapper[4808]: I1002 16:54:31.017605 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h2tkr" Oct 02 16:54:31 crc kubenswrapper[4808]: I1002 16:54:31.017660 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jz72l" podUID="21011486-b620-4b60-9b06-a7e3e003a37d" containerName="registry-server" containerID="cri-o://7246b42ea3a14518a2ebcbe0c490937720a4337719139a619153aba07c2d2169" gracePeriod=2 Oct 02 16:54:31 crc kubenswrapper[4808]: I1002 16:54:31.018060 4808 scope.go:117] "RemoveContainer" containerID="e5cfb4006a1eef69a5dd5a89cac0731ae5a7342cf2558cb5c049a15ab849b1cc" Oct 02 16:54:31 crc kubenswrapper[4808]: I1002 16:54:31.056617 4808 scope.go:117] "RemoveContainer" containerID="b0d1b1f129a158100c58b420883b6d114d282f2f1ef8acb200f051d4d4f86022" Oct 02 16:54:31 crc kubenswrapper[4808]: I1002 16:54:31.088177 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h2tkr"] Oct 02 16:54:31 crc kubenswrapper[4808]: I1002 16:54:31.109424 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h2tkr"] Oct 02 16:54:31 crc kubenswrapper[4808]: I1002 16:54:31.145434 4808 scope.go:117] "RemoveContainer" containerID="b5b6497c917a202544feed1fd2ea96bc26e272644b86a4b79cdf77ffadc0b377" Oct 02 16:54:31 crc kubenswrapper[4808]: I1002 16:54:31.406378 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fba922bf-eff6-4b8a-92bf-8a9bdce0c83f" path="/var/lib/kubelet/pods/fba922bf-eff6-4b8a-92bf-8a9bdce0c83f/volumes" Oct 02 16:54:31 crc kubenswrapper[4808]: I1002 16:54:31.461780 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jz72l" Oct 02 16:54:31 crc kubenswrapper[4808]: I1002 16:54:31.638552 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21011486-b620-4b60-9b06-a7e3e003a37d-catalog-content\") pod \"21011486-b620-4b60-9b06-a7e3e003a37d\" (UID: \"21011486-b620-4b60-9b06-a7e3e003a37d\") " Oct 02 16:54:31 crc kubenswrapper[4808]: I1002 16:54:31.638664 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8w6b\" (UniqueName: \"kubernetes.io/projected/21011486-b620-4b60-9b06-a7e3e003a37d-kube-api-access-b8w6b\") pod \"21011486-b620-4b60-9b06-a7e3e003a37d\" (UID: \"21011486-b620-4b60-9b06-a7e3e003a37d\") " Oct 02 16:54:31 crc kubenswrapper[4808]: I1002 16:54:31.638737 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21011486-b620-4b60-9b06-a7e3e003a37d-utilities\") pod \"21011486-b620-4b60-9b06-a7e3e003a37d\" (UID: \"21011486-b620-4b60-9b06-a7e3e003a37d\") " Oct 02 16:54:31 crc kubenswrapper[4808]: I1002 16:54:31.639626 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21011486-b620-4b60-9b06-a7e3e003a37d-utilities" (OuterVolumeSpecName: "utilities") pod "21011486-b620-4b60-9b06-a7e3e003a37d" (UID: "21011486-b620-4b60-9b06-a7e3e003a37d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:54:31 crc kubenswrapper[4808]: I1002 16:54:31.645405 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21011486-b620-4b60-9b06-a7e3e003a37d-kube-api-access-b8w6b" (OuterVolumeSpecName: "kube-api-access-b8w6b") pod "21011486-b620-4b60-9b06-a7e3e003a37d" (UID: "21011486-b620-4b60-9b06-a7e3e003a37d"). InnerVolumeSpecName "kube-api-access-b8w6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:54:31 crc kubenswrapper[4808]: I1002 16:54:31.711530 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21011486-b620-4b60-9b06-a7e3e003a37d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "21011486-b620-4b60-9b06-a7e3e003a37d" (UID: "21011486-b620-4b60-9b06-a7e3e003a37d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:54:31 crc kubenswrapper[4808]: I1002 16:54:31.740451 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8w6b\" (UniqueName: \"kubernetes.io/projected/21011486-b620-4b60-9b06-a7e3e003a37d-kube-api-access-b8w6b\") on node \"crc\" DevicePath \"\"" Oct 02 16:54:31 crc kubenswrapper[4808]: I1002 16:54:31.740492 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21011486-b620-4b60-9b06-a7e3e003a37d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 16:54:31 crc kubenswrapper[4808]: I1002 16:54:31.740503 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21011486-b620-4b60-9b06-a7e3e003a37d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 16:54:32 crc kubenswrapper[4808]: I1002 16:54:32.031636 4808 generic.go:334] "Generic (PLEG): container finished" podID="21011486-b620-4b60-9b06-a7e3e003a37d" containerID="7246b42ea3a14518a2ebcbe0c490937720a4337719139a619153aba07c2d2169" exitCode=0 Oct 02 16:54:32 crc kubenswrapper[4808]: I1002 16:54:32.031693 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jz72l" event={"ID":"21011486-b620-4b60-9b06-a7e3e003a37d","Type":"ContainerDied","Data":"7246b42ea3a14518a2ebcbe0c490937720a4337719139a619153aba07c2d2169"} Oct 02 16:54:32 crc kubenswrapper[4808]: I1002 16:54:32.031728 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jz72l" event={"ID":"21011486-b620-4b60-9b06-a7e3e003a37d","Type":"ContainerDied","Data":"1cc0c05dd409ef286f4d1f67c9f7404b9cd7fc13ee5caca6130138c1ff90623d"} Oct 02 16:54:32 crc kubenswrapper[4808]: I1002 16:54:32.031764 4808 scope.go:117] "RemoveContainer" containerID="7246b42ea3a14518a2ebcbe0c490937720a4337719139a619153aba07c2d2169" Oct 02 16:54:32 crc kubenswrapper[4808]: I1002 16:54:32.031830 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jz72l" Oct 02 16:54:32 crc kubenswrapper[4808]: I1002 16:54:32.059982 4808 scope.go:117] "RemoveContainer" containerID="fdd4714c2c77183454a28ee4a6fc8f9e20eb123c77353659006b7304b6568fb6" Oct 02 16:54:32 crc kubenswrapper[4808]: I1002 16:54:32.078402 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jz72l"] Oct 02 16:54:32 crc kubenswrapper[4808]: I1002 16:54:32.082191 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jz72l"] Oct 02 16:54:32 crc kubenswrapper[4808]: I1002 16:54:32.105018 4808 scope.go:117] "RemoveContainer" containerID="05fe5a1df9a5d8e80cd7c17cdda684ac4ffba3071c4fe6e86777f643b8773dc1" Oct 02 16:54:32 crc kubenswrapper[4808]: I1002 16:54:32.129068 4808 scope.go:117] "RemoveContainer" containerID="7246b42ea3a14518a2ebcbe0c490937720a4337719139a619153aba07c2d2169" Oct 02 16:54:32 crc kubenswrapper[4808]: E1002 16:54:32.129536 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7246b42ea3a14518a2ebcbe0c490937720a4337719139a619153aba07c2d2169\": container with ID starting with 7246b42ea3a14518a2ebcbe0c490937720a4337719139a619153aba07c2d2169 not found: ID does not exist" containerID="7246b42ea3a14518a2ebcbe0c490937720a4337719139a619153aba07c2d2169" Oct 02 16:54:32 crc kubenswrapper[4808]: I1002 16:54:32.129590 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7246b42ea3a14518a2ebcbe0c490937720a4337719139a619153aba07c2d2169"} err="failed to get container status \"7246b42ea3a14518a2ebcbe0c490937720a4337719139a619153aba07c2d2169\": rpc error: code = NotFound desc = could not find container \"7246b42ea3a14518a2ebcbe0c490937720a4337719139a619153aba07c2d2169\": container with ID starting with 7246b42ea3a14518a2ebcbe0c490937720a4337719139a619153aba07c2d2169 not found: ID does not exist" Oct 02 16:54:32 crc kubenswrapper[4808]: I1002 16:54:32.129626 4808 scope.go:117] "RemoveContainer" containerID="fdd4714c2c77183454a28ee4a6fc8f9e20eb123c77353659006b7304b6568fb6" Oct 02 16:54:32 crc kubenswrapper[4808]: E1002 16:54:32.129928 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdd4714c2c77183454a28ee4a6fc8f9e20eb123c77353659006b7304b6568fb6\": container with ID starting with fdd4714c2c77183454a28ee4a6fc8f9e20eb123c77353659006b7304b6568fb6 not found: ID does not exist" containerID="fdd4714c2c77183454a28ee4a6fc8f9e20eb123c77353659006b7304b6568fb6" Oct 02 16:54:32 crc kubenswrapper[4808]: I1002 16:54:32.129965 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdd4714c2c77183454a28ee4a6fc8f9e20eb123c77353659006b7304b6568fb6"} err="failed to get container status \"fdd4714c2c77183454a28ee4a6fc8f9e20eb123c77353659006b7304b6568fb6\": rpc error: code = NotFound desc = could not find container \"fdd4714c2c77183454a28ee4a6fc8f9e20eb123c77353659006b7304b6568fb6\": container with ID starting with fdd4714c2c77183454a28ee4a6fc8f9e20eb123c77353659006b7304b6568fb6 not found: ID does not exist" Oct 02 16:54:32 crc kubenswrapper[4808]: I1002 16:54:32.129991 4808 scope.go:117] "RemoveContainer" containerID="05fe5a1df9a5d8e80cd7c17cdda684ac4ffba3071c4fe6e86777f643b8773dc1" Oct 02 16:54:32 crc kubenswrapper[4808]: E1002 16:54:32.130291 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05fe5a1df9a5d8e80cd7c17cdda684ac4ffba3071c4fe6e86777f643b8773dc1\": container with ID starting with 05fe5a1df9a5d8e80cd7c17cdda684ac4ffba3071c4fe6e86777f643b8773dc1 not found: ID does not exist" containerID="05fe5a1df9a5d8e80cd7c17cdda684ac4ffba3071c4fe6e86777f643b8773dc1" Oct 02 16:54:32 crc kubenswrapper[4808]: I1002 16:54:32.130324 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05fe5a1df9a5d8e80cd7c17cdda684ac4ffba3071c4fe6e86777f643b8773dc1"} err="failed to get container status \"05fe5a1df9a5d8e80cd7c17cdda684ac4ffba3071c4fe6e86777f643b8773dc1\": rpc error: code = NotFound desc = could not find container \"05fe5a1df9a5d8e80cd7c17cdda684ac4ffba3071c4fe6e86777f643b8773dc1\": container with ID starting with 05fe5a1df9a5d8e80cd7c17cdda684ac4ffba3071c4fe6e86777f643b8773dc1 not found: ID does not exist" Oct 02 16:54:33 crc kubenswrapper[4808]: I1002 16:54:33.404402 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21011486-b620-4b60-9b06-a7e3e003a37d" path="/var/lib/kubelet/pods/21011486-b620-4b60-9b06-a7e3e003a37d/volumes" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.158545 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-mnftb"] Oct 02 16:54:40 crc kubenswrapper[4808]: E1002 16:54:40.159293 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fba922bf-eff6-4b8a-92bf-8a9bdce0c83f" containerName="registry-server" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.159308 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="fba922bf-eff6-4b8a-92bf-8a9bdce0c83f" containerName="registry-server" Oct 02 16:54:40 crc kubenswrapper[4808]: E1002 16:54:40.159327 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21011486-b620-4b60-9b06-a7e3e003a37d" containerName="registry-server" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.159335 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="21011486-b620-4b60-9b06-a7e3e003a37d" containerName="registry-server" Oct 02 16:54:40 crc kubenswrapper[4808]: E1002 16:54:40.159351 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21011486-b620-4b60-9b06-a7e3e003a37d" containerName="extract-content" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.159359 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="21011486-b620-4b60-9b06-a7e3e003a37d" containerName="extract-content" Oct 02 16:54:40 crc kubenswrapper[4808]: E1002 16:54:40.159368 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21011486-b620-4b60-9b06-a7e3e003a37d" containerName="extract-utilities" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.159376 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="21011486-b620-4b60-9b06-a7e3e003a37d" containerName="extract-utilities" Oct 02 16:54:40 crc kubenswrapper[4808]: E1002 16:54:40.159390 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fba922bf-eff6-4b8a-92bf-8a9bdce0c83f" containerName="extract-content" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.159399 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="fba922bf-eff6-4b8a-92bf-8a9bdce0c83f" containerName="extract-content" Oct 02 16:54:40 crc kubenswrapper[4808]: E1002 16:54:40.159417 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fba922bf-eff6-4b8a-92bf-8a9bdce0c83f" containerName="extract-utilities" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.159425 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="fba922bf-eff6-4b8a-92bf-8a9bdce0c83f" containerName="extract-utilities" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.159746 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="fba922bf-eff6-4b8a-92bf-8a9bdce0c83f" containerName="registry-server" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.159769 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="21011486-b620-4b60-9b06-a7e3e003a37d" containerName="registry-server" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.160487 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-mnftb" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.168877 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-gnv8z"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.170364 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gnv8z" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.176095 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-d4zs5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.176288 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-25g8p" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.216185 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-gnv8z"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.227325 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b99c7c5b-z2gb5"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.228875 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84b99c7c5b-z2gb5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.232358 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-2zhp9"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.234315 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-lxzfv" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.246034 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-2zhp9" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.254933 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-6l2k2" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.278919 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-m5tw2"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.280000 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-m5tw2" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.284572 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-pllmr" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.286725 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b99c7c5b-z2gb5"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.307942 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-mnftb"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.320272 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-2zhp9"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.320317 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-m5tw2"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.327275 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-6qnjp"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.328154 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-6qnjp" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.331516 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-w6b5g" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.339543 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-kjrph"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.340484 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-kjrph" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.346176 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.349176 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-crrgq" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.353583 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-kjrph"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.362903 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-6qnjp"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.363500 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-9cm78"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.364673 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxsgb\" (UniqueName: \"kubernetes.io/projected/6253eb77-9a32-473f-8f30-837d648e51cb-kube-api-access-hxsgb\") pod \"barbican-operator-controller-manager-6ff8b75857-mnftb\" (UID: \"6253eb77-9a32-473f-8f30-837d648e51cb\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-mnftb" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.364727 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqv48\" (UniqueName: \"kubernetes.io/projected/0193c2d7-ccc0-40bc-b4a9-e483e51803ce-kube-api-access-bqv48\") pod \"glance-operator-controller-manager-84b99c7c5b-z2gb5\" (UID: \"0193c2d7-ccc0-40bc-b4a9-e483e51803ce\") " pod="openstack-operators/glance-operator-controller-manager-84b99c7c5b-z2gb5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.365165 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb9mp\" (UniqueName: \"kubernetes.io/projected/e1981fc0-4eab-4cec-8c2a-611ab0d2af38-kube-api-access-hb9mp\") pod \"cinder-operator-controller-manager-644bddb6d8-gnv8z\" (UID: \"e1981fc0-4eab-4cec-8c2a-611ab0d2af38\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gnv8z" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.365195 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zwnk\" (UniqueName: \"kubernetes.io/projected/7546f33a-e8d7-4e15-bb6f-bd91a95d7ae9-kube-api-access-9zwnk\") pod \"designate-operator-controller-manager-84f4f7b77b-2zhp9\" (UID: \"7546f33a-e8d7-4e15-bb6f-bd91a95d7ae9\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-2zhp9" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.368610 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-9cm78" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.377105 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-gz2xs" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.377117 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.378687 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.381161 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-km6tt" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.389628 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-9cm78"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.395085 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.403719 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-wpps5"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.404606 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-wpps5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.407078 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-hn748" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.411287 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-wpps5"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.415021 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-dqz59"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.416050 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-dqz59" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.426220 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-m5vpv" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.443076 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-lz5ls"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.443980 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-lz5ls" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.445767 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-jtwck" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.449333 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-lz5ls"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.452997 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-dqz59"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.459512 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-hsmt5"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.460526 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hsmt5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.463377 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-5zkc8" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.467995 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4lp7\" (UniqueName: \"kubernetes.io/projected/4a8f2a74-9110-4f79-bc84-c68ed74b982d-kube-api-access-l4lp7\") pod \"horizon-operator-controller-manager-9f4696d94-6qnjp\" (UID: \"4a8f2a74-9110-4f79-bc84-c68ed74b982d\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-6qnjp" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.468073 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wcsz\" (UniqueName: \"kubernetes.io/projected/8771edaf-8b38-4efe-9a99-ea74be174f75-kube-api-access-6wcsz\") pod \"heat-operator-controller-manager-5d889d78cf-m5tw2\" (UID: \"8771edaf-8b38-4efe-9a99-ea74be174f75\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-m5tw2" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.468122 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r226\" (UniqueName: \"kubernetes.io/projected/e2c80909-0584-4f0c-a989-fe053b196f98-kube-api-access-5r226\") pod \"infra-operator-controller-manager-9d6c5db85-kjrph\" (UID: \"e2c80909-0584-4f0c-a989-fe053b196f98\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-kjrph" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.468157 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb9mp\" (UniqueName: \"kubernetes.io/projected/e1981fc0-4eab-4cec-8c2a-611ab0d2af38-kube-api-access-hb9mp\") pod \"cinder-operator-controller-manager-644bddb6d8-gnv8z\" (UID: \"e1981fc0-4eab-4cec-8c2a-611ab0d2af38\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gnv8z" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.468321 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zwnk\" (UniqueName: \"kubernetes.io/projected/7546f33a-e8d7-4e15-bb6f-bd91a95d7ae9-kube-api-access-9zwnk\") pod \"designate-operator-controller-manager-84f4f7b77b-2zhp9\" (UID: \"7546f33a-e8d7-4e15-bb6f-bd91a95d7ae9\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-2zhp9" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.468375 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxsgb\" (UniqueName: \"kubernetes.io/projected/6253eb77-9a32-473f-8f30-837d648e51cb-kube-api-access-hxsgb\") pod \"barbican-operator-controller-manager-6ff8b75857-mnftb\" (UID: \"6253eb77-9a32-473f-8f30-837d648e51cb\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-mnftb" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.468403 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhlw7\" (UniqueName: \"kubernetes.io/projected/4b8b017f-fc2b-4925-ab5c-001ca052ff8c-kube-api-access-hhlw7\") pod \"ironic-operator-controller-manager-5cd4858477-9cm78\" (UID: \"4b8b017f-fc2b-4925-ab5c-001ca052ff8c\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-9cm78" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.468440 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e2c80909-0584-4f0c-a989-fe053b196f98-cert\") pod \"infra-operator-controller-manager-9d6c5db85-kjrph\" (UID: \"e2c80909-0584-4f0c-a989-fe053b196f98\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-kjrph" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.468484 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqv48\" (UniqueName: \"kubernetes.io/projected/0193c2d7-ccc0-40bc-b4a9-e483e51803ce-kube-api-access-bqv48\") pod \"glance-operator-controller-manager-84b99c7c5b-z2gb5\" (UID: \"0193c2d7-ccc0-40bc-b4a9-e483e51803ce\") " pod="openstack-operators/glance-operator-controller-manager-84b99c7c5b-z2gb5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.478174 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-5tt4q"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.479212 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-5tt4q" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.484152 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-hsmt5"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.486368 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-tqlhh" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.493618 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-5tt4q"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.501834 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqv48\" (UniqueName: \"kubernetes.io/projected/0193c2d7-ccc0-40bc-b4a9-e483e51803ce-kube-api-access-bqv48\") pod \"glance-operator-controller-manager-84b99c7c5b-z2gb5\" (UID: \"0193c2d7-ccc0-40bc-b4a9-e483e51803ce\") " pod="openstack-operators/glance-operator-controller-manager-84b99c7c5b-z2gb5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.502498 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zwnk\" (UniqueName: \"kubernetes.io/projected/7546f33a-e8d7-4e15-bb6f-bd91a95d7ae9-kube-api-access-9zwnk\") pod \"designate-operator-controller-manager-84f4f7b77b-2zhp9\" (UID: \"7546f33a-e8d7-4e15-bb6f-bd91a95d7ae9\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-2zhp9" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.503870 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb9mp\" (UniqueName: \"kubernetes.io/projected/e1981fc0-4eab-4cec-8c2a-611ab0d2af38-kube-api-access-hb9mp\") pod \"cinder-operator-controller-manager-644bddb6d8-gnv8z\" (UID: \"e1981fc0-4eab-4cec-8c2a-611ab0d2af38\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gnv8z" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.517279 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxsgb\" (UniqueName: \"kubernetes.io/projected/6253eb77-9a32-473f-8f30-837d648e51cb-kube-api-access-hxsgb\") pod \"barbican-operator-controller-manager-6ff8b75857-mnftb\" (UID: \"6253eb77-9a32-473f-8f30-837d648e51cb\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-mnftb" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.522068 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-hw4dk"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.523444 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-hw4dk" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.527549 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-l8wbk" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.543971 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-zw4d5"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.545009 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-zw4d5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.546674 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.546727 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-tk746" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.564751 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84b99c7c5b-z2gb5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.565650 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-hw4dk"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.569596 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e2c80909-0584-4f0c-a989-fe053b196f98-cert\") pod \"infra-operator-controller-manager-9d6c5db85-kjrph\" (UID: \"e2c80909-0584-4f0c-a989-fe053b196f98\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-kjrph" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.569682 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4lp7\" (UniqueName: \"kubernetes.io/projected/4a8f2a74-9110-4f79-bc84-c68ed74b982d-kube-api-access-l4lp7\") pod \"horizon-operator-controller-manager-9f4696d94-6qnjp\" (UID: \"4a8f2a74-9110-4f79-bc84-c68ed74b982d\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-6qnjp" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.569715 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nhmc\" (UniqueName: \"kubernetes.io/projected/54002291-52de-4cd8-9c01-7a3bd3c2b98b-kube-api-access-4nhmc\") pod \"neutron-operator-controller-manager-849d5b9b84-lz5ls\" (UID: \"54002291-52de-4cd8-9c01-7a3bd3c2b98b\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-lz5ls" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.569749 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmxhg\" (UniqueName: \"kubernetes.io/projected/818e02a3-653a-4b36-a68b-9d4e2df0b71e-kube-api-access-hmxhg\") pod \"keystone-operator-controller-manager-65f4bcbb94-svkdd\" (UID: \"818e02a3-653a-4b36-a68b-9d4e2df0b71e\") " pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.569770 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wcsz\" (UniqueName: \"kubernetes.io/projected/8771edaf-8b38-4efe-9a99-ea74be174f75-kube-api-access-6wcsz\") pod \"heat-operator-controller-manager-5d889d78cf-m5tw2\" (UID: \"8771edaf-8b38-4efe-9a99-ea74be174f75\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-m5tw2" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.569787 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv6wr\" (UniqueName: \"kubernetes.io/projected/43dd50ad-2f9d-47a6-b685-379dd593449c-kube-api-access-gv6wr\") pod \"mariadb-operator-controller-manager-88c7-dqz59\" (UID: \"43dd50ad-2f9d-47a6-b685-379dd593449c\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-dqz59" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.569803 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kctpm\" (UniqueName: \"kubernetes.io/projected/23f6eba8-a786-4234-869f-6e44016536dd-kube-api-access-kctpm\") pod \"nova-operator-controller-manager-64cd67b5cb-hsmt5\" (UID: \"23f6eba8-a786-4234-869f-6e44016536dd\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hsmt5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.569829 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r226\" (UniqueName: \"kubernetes.io/projected/e2c80909-0584-4f0c-a989-fe053b196f98-kube-api-access-5r226\") pod \"infra-operator-controller-manager-9d6c5db85-kjrph\" (UID: \"e2c80909-0584-4f0c-a989-fe053b196f98\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-kjrph" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.569861 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qw59\" (UniqueName: \"kubernetes.io/projected/14931328-29f9-4a03-9a3b-50cb2bb861fe-kube-api-access-4qw59\") pod \"manila-operator-controller-manager-6d68dbc695-wpps5\" (UID: \"14931328-29f9-4a03-9a3b-50cb2bb861fe\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-wpps5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.569890 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhlw7\" (UniqueName: \"kubernetes.io/projected/4b8b017f-fc2b-4925-ab5c-001ca052ff8c-kube-api-access-hhlw7\") pod \"ironic-operator-controller-manager-5cd4858477-9cm78\" (UID: \"4b8b017f-fc2b-4925-ab5c-001ca052ff8c\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-9cm78" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.572771 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-zw4d5"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.587956 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-2zhp9" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.590561 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e2c80909-0584-4f0c-a989-fe053b196f98-cert\") pod \"infra-operator-controller-manager-9d6c5db85-kjrph\" (UID: \"e2c80909-0584-4f0c-a989-fe053b196f98\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-kjrph" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.594414 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhlw7\" (UniqueName: \"kubernetes.io/projected/4b8b017f-fc2b-4925-ab5c-001ca052ff8c-kube-api-access-hhlw7\") pod \"ironic-operator-controller-manager-5cd4858477-9cm78\" (UID: \"4b8b017f-fc2b-4925-ab5c-001ca052ff8c\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-9cm78" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.598031 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wcsz\" (UniqueName: \"kubernetes.io/projected/8771edaf-8b38-4efe-9a99-ea74be174f75-kube-api-access-6wcsz\") pod \"heat-operator-controller-manager-5d889d78cf-m5tw2\" (UID: \"8771edaf-8b38-4efe-9a99-ea74be174f75\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-m5tw2" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.599714 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-g7gmf"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.601680 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g7gmf" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.603577 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-2hhrj" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.604416 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4lp7\" (UniqueName: \"kubernetes.io/projected/4a8f2a74-9110-4f79-bc84-c68ed74b982d-kube-api-access-l4lp7\") pod \"horizon-operator-controller-manager-9f4696d94-6qnjp\" (UID: \"4a8f2a74-9110-4f79-bc84-c68ed74b982d\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-6qnjp" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.609091 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-8bbm9"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.611431 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r226\" (UniqueName: \"kubernetes.io/projected/e2c80909-0584-4f0c-a989-fe053b196f98-kube-api-access-5r226\") pod \"infra-operator-controller-manager-9d6c5db85-kjrph\" (UID: \"e2c80909-0584-4f0c-a989-fe053b196f98\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-kjrph" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.612612 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-8bbm9" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.614795 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-w4w6f" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.616989 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-g7gmf"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.623295 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-8bbm9"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.645020 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-6qnjp" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.657508 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-kjrph" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.670928 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/56b8aff7-4d44-41c4-b4a7-caed261ab974-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-zw4d5\" (UID: \"56b8aff7-4d44-41c4-b4a7-caed261ab974\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-zw4d5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.670968 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdj7v\" (UniqueName: \"kubernetes.io/projected/7737381d-608c-43f8-9cdf-a5612d28bd43-kube-api-access-qdj7v\") pod \"octavia-operator-controller-manager-7b787867f4-5tt4q\" (UID: \"7737381d-608c-43f8-9cdf-a5612d28bd43\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-5tt4q" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.670988 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw7sl\" (UniqueName: \"kubernetes.io/projected/56b8aff7-4d44-41c4-b4a7-caed261ab974-kube-api-access-jw7sl\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-zw4d5\" (UID: \"56b8aff7-4d44-41c4-b4a7-caed261ab974\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-zw4d5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.671064 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nhmc\" (UniqueName: \"kubernetes.io/projected/54002291-52de-4cd8-9c01-7a3bd3c2b98b-kube-api-access-4nhmc\") pod \"neutron-operator-controller-manager-849d5b9b84-lz5ls\" (UID: \"54002291-52de-4cd8-9c01-7a3bd3c2b98b\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-lz5ls" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.671099 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzl2s\" (UniqueName: \"kubernetes.io/projected/0e51874f-4bce-4f5c-8e00-99f74b1f2881-kube-api-access-bzl2s\") pod \"ovn-operator-controller-manager-9976ff44c-hw4dk\" (UID: \"0e51874f-4bce-4f5c-8e00-99f74b1f2881\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-hw4dk" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.671118 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmxhg\" (UniqueName: \"kubernetes.io/projected/818e02a3-653a-4b36-a68b-9d4e2df0b71e-kube-api-access-hmxhg\") pod \"keystone-operator-controller-manager-65f4bcbb94-svkdd\" (UID: \"818e02a3-653a-4b36-a68b-9d4e2df0b71e\") " pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.671139 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv6wr\" (UniqueName: \"kubernetes.io/projected/43dd50ad-2f9d-47a6-b685-379dd593449c-kube-api-access-gv6wr\") pod \"mariadb-operator-controller-manager-88c7-dqz59\" (UID: \"43dd50ad-2f9d-47a6-b685-379dd593449c\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-dqz59" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.671158 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kctpm\" (UniqueName: \"kubernetes.io/projected/23f6eba8-a786-4234-869f-6e44016536dd-kube-api-access-kctpm\") pod \"nova-operator-controller-manager-64cd67b5cb-hsmt5\" (UID: \"23f6eba8-a786-4234-869f-6e44016536dd\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hsmt5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.671210 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qw59\" (UniqueName: \"kubernetes.io/projected/14931328-29f9-4a03-9a3b-50cb2bb861fe-kube-api-access-4qw59\") pod \"manila-operator-controller-manager-6d68dbc695-wpps5\" (UID: \"14931328-29f9-4a03-9a3b-50cb2bb861fe\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-wpps5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.684935 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vth7d"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.686656 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vth7d" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.689860 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-9cm78" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.693597 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-jrg55" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.694456 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vth7d"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.694737 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv6wr\" (UniqueName: \"kubernetes.io/projected/43dd50ad-2f9d-47a6-b685-379dd593449c-kube-api-access-gv6wr\") pod \"mariadb-operator-controller-manager-88c7-dqz59\" (UID: \"43dd50ad-2f9d-47a6-b685-379dd593449c\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-dqz59" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.694744 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qw59\" (UniqueName: \"kubernetes.io/projected/14931328-29f9-4a03-9a3b-50cb2bb861fe-kube-api-access-4qw59\") pod \"manila-operator-controller-manager-6d68dbc695-wpps5\" (UID: \"14931328-29f9-4a03-9a3b-50cb2bb861fe\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-wpps5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.695309 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmxhg\" (UniqueName: \"kubernetes.io/projected/818e02a3-653a-4b36-a68b-9d4e2df0b71e-kube-api-access-hmxhg\") pod \"keystone-operator-controller-manager-65f4bcbb94-svkdd\" (UID: \"818e02a3-653a-4b36-a68b-9d4e2df0b71e\") " pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.695731 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kctpm\" (UniqueName: \"kubernetes.io/projected/23f6eba8-a786-4234-869f-6e44016536dd-kube-api-access-kctpm\") pod \"nova-operator-controller-manager-64cd67b5cb-hsmt5\" (UID: \"23f6eba8-a786-4234-869f-6e44016536dd\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hsmt5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.699371 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nhmc\" (UniqueName: \"kubernetes.io/projected/54002291-52de-4cd8-9c01-7a3bd3c2b98b-kube-api-access-4nhmc\") pod \"neutron-operator-controller-manager-849d5b9b84-lz5ls\" (UID: \"54002291-52de-4cd8-9c01-7a3bd3c2b98b\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-lz5ls" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.702467 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.717637 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-skv6k"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.719126 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85777745bb-skv6k" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.723569 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-4vkwm" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.726312 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-skv6k"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.733548 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-wpps5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.751342 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-dqz59" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.765065 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-lz5ls" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.770749 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-9t5j7"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.771909 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5sv9\" (UniqueName: \"kubernetes.io/projected/2a264f04-66b8-4938-ab3f-58c651166227-kube-api-access-d5sv9\") pod \"swift-operator-controller-manager-84d6b4b759-8bbm9\" (UID: \"2a264f04-66b8-4938-ab3f-58c651166227\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-8bbm9" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.771980 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/56b8aff7-4d44-41c4-b4a7-caed261ab974-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-zw4d5\" (UID: \"56b8aff7-4d44-41c4-b4a7-caed261ab974\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-zw4d5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.772003 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdj7v\" (UniqueName: \"kubernetes.io/projected/7737381d-608c-43f8-9cdf-a5612d28bd43-kube-api-access-qdj7v\") pod \"octavia-operator-controller-manager-7b787867f4-5tt4q\" (UID: \"7737381d-608c-43f8-9cdf-a5612d28bd43\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-5tt4q" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.772024 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw7sl\" (UniqueName: \"kubernetes.io/projected/56b8aff7-4d44-41c4-b4a7-caed261ab974-kube-api-access-jw7sl\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-zw4d5\" (UID: \"56b8aff7-4d44-41c4-b4a7-caed261ab974\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-zw4d5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.772079 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzl2s\" (UniqueName: \"kubernetes.io/projected/0e51874f-4bce-4f5c-8e00-99f74b1f2881-kube-api-access-bzl2s\") pod \"ovn-operator-controller-manager-9976ff44c-hw4dk\" (UID: \"0e51874f-4bce-4f5c-8e00-99f74b1f2881\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-hw4dk" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.772118 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqs7v\" (UniqueName: \"kubernetes.io/projected/0cf6cc81-9904-4904-bf45-a4d3e800231b-kube-api-access-kqs7v\") pod \"placement-operator-controller-manager-589c58c6c-g7gmf\" (UID: \"0cf6cc81-9904-4904-bf45-a4d3e800231b\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g7gmf" Oct 02 16:54:40 crc kubenswrapper[4808]: E1002 16:54:40.772156 4808 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 16:54:40 crc kubenswrapper[4808]: E1002 16:54:40.772216 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/56b8aff7-4d44-41c4-b4a7-caed261ab974-cert podName:56b8aff7-4d44-41c4-b4a7-caed261ab974 nodeName:}" failed. No retries permitted until 2025-10-02 16:54:41.272197003 +0000 UTC m=+848.597726003 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/56b8aff7-4d44-41c4-b4a7-caed261ab974-cert") pod "openstack-baremetal-operator-controller-manager-5869cb545-zw4d5" (UID: "56b8aff7-4d44-41c4-b4a7-caed261ab974") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.772823 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9t5j7" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.774988 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-9t5j7"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.787127 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-d72zf" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.787224 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-mnftb" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.788015 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hsmt5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.788061 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gnv8z" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.797225 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdj7v\" (UniqueName: \"kubernetes.io/projected/7737381d-608c-43f8-9cdf-a5612d28bd43-kube-api-access-qdj7v\") pod \"octavia-operator-controller-manager-7b787867f4-5tt4q\" (UID: \"7737381d-608c-43f8-9cdf-a5612d28bd43\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-5tt4q" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.802783 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzl2s\" (UniqueName: \"kubernetes.io/projected/0e51874f-4bce-4f5c-8e00-99f74b1f2881-kube-api-access-bzl2s\") pod \"ovn-operator-controller-manager-9976ff44c-hw4dk\" (UID: \"0e51874f-4bce-4f5c-8e00-99f74b1f2881\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-hw4dk" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.804940 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw7sl\" (UniqueName: \"kubernetes.io/projected/56b8aff7-4d44-41c4-b4a7-caed261ab974-kube-api-access-jw7sl\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-zw4d5\" (UID: \"56b8aff7-4d44-41c4-b4a7-caed261ab974\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-zw4d5" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.838797 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-5tt4q" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.857612 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-hw4dk" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.870567 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6b57c44d8-px47q"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.871642 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6b57c44d8-px47q" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.873578 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59flp\" (UniqueName: \"kubernetes.io/projected/d2828482-d7e8-4e5f-985d-197d614d8d15-kube-api-access-59flp\") pod \"watcher-operator-controller-manager-6b9957f54f-9t5j7\" (UID: \"d2828482-d7e8-4e5f-985d-197d614d8d15\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9t5j7" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.873627 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqs7v\" (UniqueName: \"kubernetes.io/projected/0cf6cc81-9904-4904-bf45-a4d3e800231b-kube-api-access-kqs7v\") pod \"placement-operator-controller-manager-589c58c6c-g7gmf\" (UID: \"0cf6cc81-9904-4904-bf45-a4d3e800231b\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g7gmf" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.873653 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5sv9\" (UniqueName: \"kubernetes.io/projected/2a264f04-66b8-4938-ab3f-58c651166227-kube-api-access-d5sv9\") pod \"swift-operator-controller-manager-84d6b4b759-8bbm9\" (UID: \"2a264f04-66b8-4938-ab3f-58c651166227\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-8bbm9" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.873676 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbxfk\" (UniqueName: \"kubernetes.io/projected/1ec1bdf1-04ec-4b70-8e58-5df6541b021f-kube-api-access-kbxfk\") pod \"test-operator-controller-manager-85777745bb-skv6k\" (UID: \"1ec1bdf1-04ec-4b70-8e58-5df6541b021f\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-skv6k" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.873724 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj8bx\" (UniqueName: \"kubernetes.io/projected/8b68ec2c-6583-482b-a9f4-d75fab859c95-kube-api-access-cj8bx\") pod \"telemetry-operator-controller-manager-b8d54b5d7-vth7d\" (UID: \"8b68ec2c-6583-482b-a9f4-d75fab859c95\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vth7d" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.876348 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.876940 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-pnw7j" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.882455 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6b57c44d8-px47q"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.892157 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqs7v\" (UniqueName: \"kubernetes.io/projected/0cf6cc81-9904-4904-bf45-a4d3e800231b-kube-api-access-kqs7v\") pod \"placement-operator-controller-manager-589c58c6c-g7gmf\" (UID: \"0cf6cc81-9904-4904-bf45-a4d3e800231b\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g7gmf" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.897581 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-m5tw2" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.899543 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5sv9\" (UniqueName: \"kubernetes.io/projected/2a264f04-66b8-4938-ab3f-58c651166227-kube-api-access-d5sv9\") pod \"swift-operator-controller-manager-84d6b4b759-8bbm9\" (UID: \"2a264f04-66b8-4938-ab3f-58c651166227\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-8bbm9" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.962193 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-wv9sn"] Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.963065 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-wv9sn" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.965601 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g7gmf" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.966046 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-vkb4c" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.973503 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-8bbm9" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.974798 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj8bx\" (UniqueName: \"kubernetes.io/projected/8b68ec2c-6583-482b-a9f4-d75fab859c95-kube-api-access-cj8bx\") pod \"telemetry-operator-controller-manager-b8d54b5d7-vth7d\" (UID: \"8b68ec2c-6583-482b-a9f4-d75fab859c95\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vth7d" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.974839 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59flp\" (UniqueName: \"kubernetes.io/projected/d2828482-d7e8-4e5f-985d-197d614d8d15-kube-api-access-59flp\") pod \"watcher-operator-controller-manager-6b9957f54f-9t5j7\" (UID: \"d2828482-d7e8-4e5f-985d-197d614d8d15\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9t5j7" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.974865 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4b8037b3-28b8-4492-b4a7-7ed5413df620-cert\") pod \"openstack-operator-controller-manager-6b57c44d8-px47q\" (UID: \"4b8037b3-28b8-4492-b4a7-7ed5413df620\") " pod="openstack-operators/openstack-operator-controller-manager-6b57c44d8-px47q" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.974963 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbxfk\" (UniqueName: \"kubernetes.io/projected/1ec1bdf1-04ec-4b70-8e58-5df6541b021f-kube-api-access-kbxfk\") pod \"test-operator-controller-manager-85777745bb-skv6k\" (UID: \"1ec1bdf1-04ec-4b70-8e58-5df6541b021f\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-skv6k" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.975010 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wckpx\" (UniqueName: \"kubernetes.io/projected/4b8037b3-28b8-4492-b4a7-7ed5413df620-kube-api-access-wckpx\") pod \"openstack-operator-controller-manager-6b57c44d8-px47q\" (UID: \"4b8037b3-28b8-4492-b4a7-7ed5413df620\") " pod="openstack-operators/openstack-operator-controller-manager-6b57c44d8-px47q" Oct 02 16:54:40 crc kubenswrapper[4808]: I1002 16:54:40.977737 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-wv9sn"] Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.006820 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59flp\" (UniqueName: \"kubernetes.io/projected/d2828482-d7e8-4e5f-985d-197d614d8d15-kube-api-access-59flp\") pod \"watcher-operator-controller-manager-6b9957f54f-9t5j7\" (UID: \"d2828482-d7e8-4e5f-985d-197d614d8d15\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9t5j7" Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.011772 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj8bx\" (UniqueName: \"kubernetes.io/projected/8b68ec2c-6583-482b-a9f4-d75fab859c95-kube-api-access-cj8bx\") pod \"telemetry-operator-controller-manager-b8d54b5d7-vth7d\" (UID: \"8b68ec2c-6583-482b-a9f4-d75fab859c95\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vth7d" Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.014792 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vth7d" Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.025591 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbxfk\" (UniqueName: \"kubernetes.io/projected/1ec1bdf1-04ec-4b70-8e58-5df6541b021f-kube-api-access-kbxfk\") pod \"test-operator-controller-manager-85777745bb-skv6k\" (UID: \"1ec1bdf1-04ec-4b70-8e58-5df6541b021f\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-skv6k" Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.056092 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85777745bb-skv6k" Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.075860 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wckpx\" (UniqueName: \"kubernetes.io/projected/4b8037b3-28b8-4492-b4a7-7ed5413df620-kube-api-access-wckpx\") pod \"openstack-operator-controller-manager-6b57c44d8-px47q\" (UID: \"4b8037b3-28b8-4492-b4a7-7ed5413df620\") " pod="openstack-operators/openstack-operator-controller-manager-6b57c44d8-px47q" Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.075915 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4b8037b3-28b8-4492-b4a7-7ed5413df620-cert\") pod \"openstack-operator-controller-manager-6b57c44d8-px47q\" (UID: \"4b8037b3-28b8-4492-b4a7-7ed5413df620\") " pod="openstack-operators/openstack-operator-controller-manager-6b57c44d8-px47q" Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.075991 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52vsw\" (UniqueName: \"kubernetes.io/projected/ccfa7a5b-15b3-4fd1-9a36-f04d894a838f-kube-api-access-52vsw\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-wv9sn\" (UID: \"ccfa7a5b-15b3-4fd1-9a36-f04d894a838f\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-wv9sn" Oct 02 16:54:41 crc kubenswrapper[4808]: E1002 16:54:41.076133 4808 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 02 16:54:41 crc kubenswrapper[4808]: E1002 16:54:41.076210 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4b8037b3-28b8-4492-b4a7-7ed5413df620-cert podName:4b8037b3-28b8-4492-b4a7-7ed5413df620 nodeName:}" failed. No retries permitted until 2025-10-02 16:54:41.576189167 +0000 UTC m=+848.901718167 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4b8037b3-28b8-4492-b4a7-7ed5413df620-cert") pod "openstack-operator-controller-manager-6b57c44d8-px47q" (UID: "4b8037b3-28b8-4492-b4a7-7ed5413df620") : secret "webhook-server-cert" not found Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.094268 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wckpx\" (UniqueName: \"kubernetes.io/projected/4b8037b3-28b8-4492-b4a7-7ed5413df620-kube-api-access-wckpx\") pod \"openstack-operator-controller-manager-6b57c44d8-px47q\" (UID: \"4b8037b3-28b8-4492-b4a7-7ed5413df620\") " pod="openstack-operators/openstack-operator-controller-manager-6b57c44d8-px47q" Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.125518 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9t5j7" Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.175092 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b99c7c5b-z2gb5"] Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.177477 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52vsw\" (UniqueName: \"kubernetes.io/projected/ccfa7a5b-15b3-4fd1-9a36-f04d894a838f-kube-api-access-52vsw\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-wv9sn\" (UID: \"ccfa7a5b-15b3-4fd1-9a36-f04d894a838f\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-wv9sn" Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.178812 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-2zhp9"] Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.194774 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52vsw\" (UniqueName: \"kubernetes.io/projected/ccfa7a5b-15b3-4fd1-9a36-f04d894a838f-kube-api-access-52vsw\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-wv9sn\" (UID: \"ccfa7a5b-15b3-4fd1-9a36-f04d894a838f\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-wv9sn" Oct 02 16:54:41 crc kubenswrapper[4808]: W1002 16:54:41.239191 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0193c2d7_ccc0_40bc_b4a9_e483e51803ce.slice/crio-f5b15cd975a191bdde409eeff1d94f12c2e2620bbc2d065ec67c5f334e7d41c1 WatchSource:0}: Error finding container f5b15cd975a191bdde409eeff1d94f12c2e2620bbc2d065ec67c5f334e7d41c1: Status 404 returned error can't find the container with id f5b15cd975a191bdde409eeff1d94f12c2e2620bbc2d065ec67c5f334e7d41c1 Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.279567 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/56b8aff7-4d44-41c4-b4a7-caed261ab974-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-zw4d5\" (UID: \"56b8aff7-4d44-41c4-b4a7-caed261ab974\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-zw4d5" Oct 02 16:54:41 crc kubenswrapper[4808]: E1002 16:54:41.279829 4808 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 16:54:41 crc kubenswrapper[4808]: E1002 16:54:41.279894 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/56b8aff7-4d44-41c4-b4a7-caed261ab974-cert podName:56b8aff7-4d44-41c4-b4a7-caed261ab974 nodeName:}" failed. No retries permitted until 2025-10-02 16:54:42.279878094 +0000 UTC m=+849.605407094 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/56b8aff7-4d44-41c4-b4a7-caed261ab974-cert") pod "openstack-baremetal-operator-controller-manager-5869cb545-zw4d5" (UID: "56b8aff7-4d44-41c4-b4a7-caed261ab974") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.344435 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-wv9sn" Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.363354 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-6qnjp"] Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.587645 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4b8037b3-28b8-4492-b4a7-7ed5413df620-cert\") pod \"openstack-operator-controller-manager-6b57c44d8-px47q\" (UID: \"4b8037b3-28b8-4492-b4a7-7ed5413df620\") " pod="openstack-operators/openstack-operator-controller-manager-6b57c44d8-px47q" Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.593979 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4b8037b3-28b8-4492-b4a7-7ed5413df620-cert\") pod \"openstack-operator-controller-manager-6b57c44d8-px47q\" (UID: \"4b8037b3-28b8-4492-b4a7-7ed5413df620\") " pod="openstack-operators/openstack-operator-controller-manager-6b57c44d8-px47q" Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.735474 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd"] Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.751924 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-kjrph"] Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.763289 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-dqz59"] Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.770728 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-hsmt5"] Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.778497 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-9cm78"] Oct 02 16:54:41 crc kubenswrapper[4808]: W1002 16:54:41.782348 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod818e02a3_653a_4b36_a68b_9d4e2df0b71e.slice/crio-010b1981107b12f794c7a775f1055c58409ee0c9971eaa56a4e518301dd33b15 WatchSource:0}: Error finding container 010b1981107b12f794c7a775f1055c58409ee0c9971eaa56a4e518301dd33b15: Status 404 returned error can't find the container with id 010b1981107b12f794c7a775f1055c58409ee0c9971eaa56a4e518301dd33b15 Oct 02 16:54:41 crc kubenswrapper[4808]: W1002 16:54:41.784709 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b8b017f_fc2b_4925_ab5c_001ca052ff8c.slice/crio-bf14d73ef429b212bf6e87e087e595477dc42f9d4f1b5a4dadaefdab061a94e5 WatchSource:0}: Error finding container bf14d73ef429b212bf6e87e087e595477dc42f9d4f1b5a4dadaefdab061a94e5: Status 404 returned error can't find the container with id bf14d73ef429b212bf6e87e087e595477dc42f9d4f1b5a4dadaefdab061a94e5 Oct 02 16:54:41 crc kubenswrapper[4808]: W1002 16:54:41.790729 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43dd50ad_2f9d_47a6_b685_379dd593449c.slice/crio-36a312530a2546abc7802be8be2ab62a517c1e7e777a2e0f95ee4b46cf3df95b WatchSource:0}: Error finding container 36a312530a2546abc7802be8be2ab62a517c1e7e777a2e0f95ee4b46cf3df95b: Status 404 returned error can't find the container with id 36a312530a2546abc7802be8be2ab62a517c1e7e777a2e0f95ee4b46cf3df95b Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.806550 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6b57c44d8-px47q" Oct 02 16:54:41 crc kubenswrapper[4808]: I1002 16:54:41.934164 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-lz5ls"] Oct 02 16:54:41 crc kubenswrapper[4808]: W1002 16:54:41.941522 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54002291_52de_4cd8_9c01_7a3bd3c2b98b.slice/crio-f47551d53154a1fddbf2b9d472c5a58a78211d65397ee3a2eb9199b086e323e4 WatchSource:0}: Error finding container f47551d53154a1fddbf2b9d472c5a58a78211d65397ee3a2eb9199b086e323e4: Status 404 returned error can't find the container with id f47551d53154a1fddbf2b9d472c5a58a78211d65397ee3a2eb9199b086e323e4 Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.136415 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-6qnjp" event={"ID":"4a8f2a74-9110-4f79-bc84-c68ed74b982d","Type":"ContainerStarted","Data":"79862d6c037c06133b87923b84abea2f09767d5c2076067b0ec9a18e52e211d0"} Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.144659 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-lz5ls" event={"ID":"54002291-52de-4cd8-9c01-7a3bd3c2b98b","Type":"ContainerStarted","Data":"f47551d53154a1fddbf2b9d472c5a58a78211d65397ee3a2eb9199b086e323e4"} Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.145947 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-2zhp9" event={"ID":"7546f33a-e8d7-4e15-bb6f-bd91a95d7ae9","Type":"ContainerStarted","Data":"e0f1ba1ed5a859c056aa400a086883d10b47e832b08f0ea0a40d142745b48f92"} Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.151283 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-kjrph" event={"ID":"e2c80909-0584-4f0c-a989-fe053b196f98","Type":"ContainerStarted","Data":"8250ae2fb0b9af9e811d6722641821eb741780ddeb31d979a2748cee6c186932"} Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.152524 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hsmt5" event={"ID":"23f6eba8-a786-4234-869f-6e44016536dd","Type":"ContainerStarted","Data":"518a8bdf2b79b15d0bdb464ba7bc20afe31125b96cefc6a538d02bb834860c01"} Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.154097 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-dqz59" event={"ID":"43dd50ad-2f9d-47a6-b685-379dd593449c","Type":"ContainerStarted","Data":"36a312530a2546abc7802be8be2ab62a517c1e7e777a2e0f95ee4b46cf3df95b"} Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.156013 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b99c7c5b-z2gb5" event={"ID":"0193c2d7-ccc0-40bc-b4a9-e483e51803ce","Type":"ContainerStarted","Data":"f5b15cd975a191bdde409eeff1d94f12c2e2620bbc2d065ec67c5f334e7d41c1"} Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.157176 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" event={"ID":"818e02a3-653a-4b36-a68b-9d4e2df0b71e","Type":"ContainerStarted","Data":"010b1981107b12f794c7a775f1055c58409ee0c9971eaa56a4e518301dd33b15"} Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.159661 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-9cm78" event={"ID":"4b8b017f-fc2b-4925-ab5c-001ca052ff8c","Type":"ContainerStarted","Data":"bf14d73ef429b212bf6e87e087e595477dc42f9d4f1b5a4dadaefdab061a94e5"} Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.182749 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-gnv8z"] Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.190344 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-9t5j7"] Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.201916 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-hw4dk"] Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.208347 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-5tt4q"] Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.220159 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-mnftb"] Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.232168 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-skv6k"] Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.245452 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-m5tw2"] Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.245500 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-wpps5"] Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.254306 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-8bbm9"] Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.257885 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-g7gmf"] Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.276291 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-wv9sn"] Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.295340 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vth7d"] Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.297802 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/56b8aff7-4d44-41c4-b4a7-caed261ab974-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-zw4d5\" (UID: \"56b8aff7-4d44-41c4-b4a7-caed261ab974\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-zw4d5" Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.304923 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/56b8aff7-4d44-41c4-b4a7-caed261ab974-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-zw4d5\" (UID: \"56b8aff7-4d44-41c4-b4a7-caed261ab974\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-zw4d5" Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.321898 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6b57c44d8-px47q"] Oct 02 16:54:42 crc kubenswrapper[4808]: I1002 16:54:42.367619 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-zw4d5" Oct 02 16:54:43 crc kubenswrapper[4808]: W1002 16:54:43.069343 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b68ec2c_6583_482b_a9f4_d75fab859c95.slice/crio-1794b8d32886c995033a0e757a25bd5a71f272b2b99bcc81878a46a084659941 WatchSource:0}: Error finding container 1794b8d32886c995033a0e757a25bd5a71f272b2b99bcc81878a46a084659941: Status 404 returned error can't find the container with id 1794b8d32886c995033a0e757a25bd5a71f272b2b99bcc81878a46a084659941 Oct 02 16:54:43 crc kubenswrapper[4808]: W1002 16:54:43.069502 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podccfa7a5b_15b3_4fd1_9a36_f04d894a838f.slice/crio-068cf99a419827061680267502dd7ade69f49e8497be32ea5c532b7be3a5a399 WatchSource:0}: Error finding container 068cf99a419827061680267502dd7ade69f49e8497be32ea5c532b7be3a5a399: Status 404 returned error can't find the container with id 068cf99a419827061680267502dd7ade69f49e8497be32ea5c532b7be3a5a399 Oct 02 16:54:43 crc kubenswrapper[4808]: I1002 16:54:43.072582 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zs4xs"] Oct 02 16:54:43 crc kubenswrapper[4808]: I1002 16:54:43.076953 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zs4xs" Oct 02 16:54:43 crc kubenswrapper[4808]: I1002 16:54:43.091986 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zs4xs"] Oct 02 16:54:43 crc kubenswrapper[4808]: I1002 16:54:43.165372 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vth7d" event={"ID":"8b68ec2c-6583-482b-a9f4-d75fab859c95","Type":"ContainerStarted","Data":"1794b8d32886c995033a0e757a25bd5a71f272b2b99bcc81878a46a084659941"} Oct 02 16:54:43 crc kubenswrapper[4808]: I1002 16:54:43.166724 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-wv9sn" event={"ID":"ccfa7a5b-15b3-4fd1-9a36-f04d894a838f","Type":"ContainerStarted","Data":"068cf99a419827061680267502dd7ade69f49e8497be32ea5c532b7be3a5a399"} Oct 02 16:54:43 crc kubenswrapper[4808]: I1002 16:54:43.173209 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-wpps5" event={"ID":"14931328-29f9-4a03-9a3b-50cb2bb861fe","Type":"ContainerStarted","Data":"e053d3d6977c70f616f6910ad9d9cb78b1d0195d73718e4b4543abe301b27f1d"} Oct 02 16:54:43 crc kubenswrapper[4808]: I1002 16:54:43.228826 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx5tj\" (UniqueName: \"kubernetes.io/projected/98a4a627-67dd-4dac-a7a4-cdfc4baaa057-kube-api-access-zx5tj\") pod \"certified-operators-zs4xs\" (UID: \"98a4a627-67dd-4dac-a7a4-cdfc4baaa057\") " pod="openshift-marketplace/certified-operators-zs4xs" Oct 02 16:54:43 crc kubenswrapper[4808]: I1002 16:54:43.228929 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98a4a627-67dd-4dac-a7a4-cdfc4baaa057-utilities\") pod \"certified-operators-zs4xs\" (UID: \"98a4a627-67dd-4dac-a7a4-cdfc4baaa057\") " pod="openshift-marketplace/certified-operators-zs4xs" Oct 02 16:54:43 crc kubenswrapper[4808]: I1002 16:54:43.228953 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98a4a627-67dd-4dac-a7a4-cdfc4baaa057-catalog-content\") pod \"certified-operators-zs4xs\" (UID: \"98a4a627-67dd-4dac-a7a4-cdfc4baaa057\") " pod="openshift-marketplace/certified-operators-zs4xs" Oct 02 16:54:43 crc kubenswrapper[4808]: I1002 16:54:43.330991 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98a4a627-67dd-4dac-a7a4-cdfc4baaa057-utilities\") pod \"certified-operators-zs4xs\" (UID: \"98a4a627-67dd-4dac-a7a4-cdfc4baaa057\") " pod="openshift-marketplace/certified-operators-zs4xs" Oct 02 16:54:43 crc kubenswrapper[4808]: I1002 16:54:43.331052 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98a4a627-67dd-4dac-a7a4-cdfc4baaa057-catalog-content\") pod \"certified-operators-zs4xs\" (UID: \"98a4a627-67dd-4dac-a7a4-cdfc4baaa057\") " pod="openshift-marketplace/certified-operators-zs4xs" Oct 02 16:54:43 crc kubenswrapper[4808]: I1002 16:54:43.331129 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx5tj\" (UniqueName: \"kubernetes.io/projected/98a4a627-67dd-4dac-a7a4-cdfc4baaa057-kube-api-access-zx5tj\") pod \"certified-operators-zs4xs\" (UID: \"98a4a627-67dd-4dac-a7a4-cdfc4baaa057\") " pod="openshift-marketplace/certified-operators-zs4xs" Oct 02 16:54:43 crc kubenswrapper[4808]: I1002 16:54:43.331524 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98a4a627-67dd-4dac-a7a4-cdfc4baaa057-utilities\") pod \"certified-operators-zs4xs\" (UID: \"98a4a627-67dd-4dac-a7a4-cdfc4baaa057\") " pod="openshift-marketplace/certified-operators-zs4xs" Oct 02 16:54:43 crc kubenswrapper[4808]: I1002 16:54:43.331860 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98a4a627-67dd-4dac-a7a4-cdfc4baaa057-catalog-content\") pod \"certified-operators-zs4xs\" (UID: \"98a4a627-67dd-4dac-a7a4-cdfc4baaa057\") " pod="openshift-marketplace/certified-operators-zs4xs" Oct 02 16:54:43 crc kubenswrapper[4808]: I1002 16:54:43.373194 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx5tj\" (UniqueName: \"kubernetes.io/projected/98a4a627-67dd-4dac-a7a4-cdfc4baaa057-kube-api-access-zx5tj\") pod \"certified-operators-zs4xs\" (UID: \"98a4a627-67dd-4dac-a7a4-cdfc4baaa057\") " pod="openshift-marketplace/certified-operators-zs4xs" Oct 02 16:54:43 crc kubenswrapper[4808]: I1002 16:54:43.434049 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zs4xs" Oct 02 16:54:44 crc kubenswrapper[4808]: W1002 16:54:44.368973 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1981fc0_4eab_4cec_8c2a_611ab0d2af38.slice/crio-5cd024754bb8bcc1e50c1a05cc6dac10b951f7a79dc4d33f4aaab14892f8d1da WatchSource:0}: Error finding container 5cd024754bb8bcc1e50c1a05cc6dac10b951f7a79dc4d33f4aaab14892f8d1da: Status 404 returned error can't find the container with id 5cd024754bb8bcc1e50c1a05cc6dac10b951f7a79dc4d33f4aaab14892f8d1da Oct 02 16:54:44 crc kubenswrapper[4808]: W1002 16:54:44.796565 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b8037b3_28b8_4492_b4a7_7ed5413df620.slice/crio-38b4eef97b8b5739c550973bd74d9316dc8007394317cb9329b0668513bd3a67 WatchSource:0}: Error finding container 38b4eef97b8b5739c550973bd74d9316dc8007394317cb9329b0668513bd3a67: Status 404 returned error can't find the container with id 38b4eef97b8b5739c550973bd74d9316dc8007394317cb9329b0668513bd3a67 Oct 02 16:54:45 crc kubenswrapper[4808]: I1002 16:54:45.184339 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6b57c44d8-px47q" event={"ID":"4b8037b3-28b8-4492-b4a7-7ed5413df620","Type":"ContainerStarted","Data":"38b4eef97b8b5739c550973bd74d9316dc8007394317cb9329b0668513bd3a67"} Oct 02 16:54:45 crc kubenswrapper[4808]: I1002 16:54:45.185552 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gnv8z" event={"ID":"e1981fc0-4eab-4cec-8c2a-611ab0d2af38","Type":"ContainerStarted","Data":"5cd024754bb8bcc1e50c1a05cc6dac10b951f7a79dc4d33f4aaab14892f8d1da"} Oct 02 16:54:45 crc kubenswrapper[4808]: I1002 16:54:45.186557 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-skv6k" event={"ID":"1ec1bdf1-04ec-4b70-8e58-5df6541b021f","Type":"ContainerStarted","Data":"68fc282eb552c6f463ae374f21fdf2ba40b672d649927033099cd4bc8a6f5710"} Oct 02 16:54:45 crc kubenswrapper[4808]: I1002 16:54:45.187534 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-5tt4q" event={"ID":"7737381d-608c-43f8-9cdf-a5612d28bd43","Type":"ContainerStarted","Data":"0ee545f223fd5ba6780547e894ee6c61f2fb0ddf618fa1dd93f8dad73d025e4c"} Oct 02 16:54:45 crc kubenswrapper[4808]: W1002 16:54:45.359322 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0cf6cc81_9904_4904_bf45_a4d3e800231b.slice/crio-6baf97988df046c8d98edf65cff522c9aec06dd7c113142a25e84fa347356ca1 WatchSource:0}: Error finding container 6baf97988df046c8d98edf65cff522c9aec06dd7c113142a25e84fa347356ca1: Status 404 returned error can't find the container with id 6baf97988df046c8d98edf65cff522c9aec06dd7c113142a25e84fa347356ca1 Oct 02 16:54:45 crc kubenswrapper[4808]: W1002 16:54:45.364320 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e51874f_4bce_4f5c_8e00_99f74b1f2881.slice/crio-c7731810fbed1a2733370d53b55b23944089a443d516ea1436a64435b61691f1 WatchSource:0}: Error finding container c7731810fbed1a2733370d53b55b23944089a443d516ea1436a64435b61691f1: Status 404 returned error can't find the container with id c7731810fbed1a2733370d53b55b23944089a443d516ea1436a64435b61691f1 Oct 02 16:54:45 crc kubenswrapper[4808]: W1002 16:54:45.382646 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2828482_d7e8_4e5f_985d_197d614d8d15.slice/crio-99575bf29c884d9939584e3152032c8dbba8b37d8fca0ced4cb1db291a73ab0d WatchSource:0}: Error finding container 99575bf29c884d9939584e3152032c8dbba8b37d8fca0ced4cb1db291a73ab0d: Status 404 returned error can't find the container with id 99575bf29c884d9939584e3152032c8dbba8b37d8fca0ced4cb1db291a73ab0d Oct 02 16:54:46 crc kubenswrapper[4808]: I1002 16:54:46.194818 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-8bbm9" event={"ID":"2a264f04-66b8-4938-ab3f-58c651166227","Type":"ContainerStarted","Data":"68e970b6b3000f53ab24ca8a203318cdc3716557b2b984a4f6b737f4a8024cac"} Oct 02 16:54:46 crc kubenswrapper[4808]: I1002 16:54:46.196268 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-hw4dk" event={"ID":"0e51874f-4bce-4f5c-8e00-99f74b1f2881","Type":"ContainerStarted","Data":"c7731810fbed1a2733370d53b55b23944089a443d516ea1436a64435b61691f1"} Oct 02 16:54:46 crc kubenswrapper[4808]: I1002 16:54:46.197654 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-m5tw2" event={"ID":"8771edaf-8b38-4efe-9a99-ea74be174f75","Type":"ContainerStarted","Data":"62de92bee6f60aa72531d9b5044993b915b41fb30c305b69ce423f5b4113360d"} Oct 02 16:54:46 crc kubenswrapper[4808]: I1002 16:54:46.199664 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9t5j7" event={"ID":"d2828482-d7e8-4e5f-985d-197d614d8d15","Type":"ContainerStarted","Data":"99575bf29c884d9939584e3152032c8dbba8b37d8fca0ced4cb1db291a73ab0d"} Oct 02 16:54:46 crc kubenswrapper[4808]: I1002 16:54:46.200528 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-mnftb" event={"ID":"6253eb77-9a32-473f-8f30-837d648e51cb","Type":"ContainerStarted","Data":"d1609b8998cb6d43b778e14aef54e10067b61085cc6a852781ac99ffffcdc7b0"} Oct 02 16:54:46 crc kubenswrapper[4808]: I1002 16:54:46.201827 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g7gmf" event={"ID":"0cf6cc81-9904-4904-bf45-a4d3e800231b","Type":"ContainerStarted","Data":"6baf97988df046c8d98edf65cff522c9aec06dd7c113142a25e84fa347356ca1"} Oct 02 16:54:48 crc kubenswrapper[4808]: I1002 16:54:48.536143 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-zw4d5"] Oct 02 16:54:48 crc kubenswrapper[4808]: I1002 16:54:48.739885 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 16:54:48 crc kubenswrapper[4808]: I1002 16:54:48.739971 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 16:54:48 crc kubenswrapper[4808]: I1002 16:54:48.740035 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 16:54:48 crc kubenswrapper[4808]: I1002 16:54:48.741046 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8d253701a5ddc087ee8a36b225b0284b63257d65500bbf63a65e988cba6e1fa1"} pod="openshift-machine-config-operator/machine-config-daemon-7z66r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 16:54:48 crc kubenswrapper[4808]: I1002 16:54:48.741139 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" containerID="cri-o://8d253701a5ddc087ee8a36b225b0284b63257d65500bbf63a65e988cba6e1fa1" gracePeriod=600 Oct 02 16:54:49 crc kubenswrapper[4808]: I1002 16:54:49.225151 4808 generic.go:334] "Generic (PLEG): container finished" podID="f83ce425-101d-4489-94a4-5c256eb29328" containerID="8d253701a5ddc087ee8a36b225b0284b63257d65500bbf63a65e988cba6e1fa1" exitCode=0 Oct 02 16:54:49 crc kubenswrapper[4808]: I1002 16:54:49.225404 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" event={"ID":"f83ce425-101d-4489-94a4-5c256eb29328","Type":"ContainerDied","Data":"8d253701a5ddc087ee8a36b225b0284b63257d65500bbf63a65e988cba6e1fa1"} Oct 02 16:54:49 crc kubenswrapper[4808]: I1002 16:54:49.226740 4808 scope.go:117] "RemoveContainer" containerID="a9406662e09c7be4b59f910ebd57fef50f5b64e0992e5ddee0ca92a0e3d8521a" Oct 02 16:54:54 crc kubenswrapper[4808]: I1002 16:54:54.262935 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-zw4d5" event={"ID":"56b8aff7-4d44-41c4-b4a7-caed261ab974","Type":"ContainerStarted","Data":"a92118c116abe051a64cbc715e980f809d6f2fd67f5cfee92d3d277c7cd6c5f7"} Oct 02 16:55:02 crc kubenswrapper[4808]: I1002 16:55:02.688836 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zs4xs"] Oct 02 16:55:02 crc kubenswrapper[4808]: W1002 16:55:02.746928 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98a4a627_67dd_4dac_a7a4_cdfc4baaa057.slice/crio-cc50038a21de7cc2d3c6e531d89382b0703488fea5a2aaefa7d1d47f971b9279 WatchSource:0}: Error finding container cc50038a21de7cc2d3c6e531d89382b0703488fea5a2aaefa7d1d47f971b9279: Status 404 returned error can't find the container with id cc50038a21de7cc2d3c6e531d89382b0703488fea5a2aaefa7d1d47f971b9279 Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.369857 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" event={"ID":"818e02a3-653a-4b36-a68b-9d4e2df0b71e","Type":"ContainerStarted","Data":"72dfceead77ed393455b2e7fbcb7b164af0fc5382ab0ba62ad1f243b06b38645"} Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.371849 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-9cm78" event={"ID":"4b8b017f-fc2b-4925-ab5c-001ca052ff8c","Type":"ContainerStarted","Data":"bf34ef5cc187f17a43512d39608acf176475a3c88273b9555d8f4d3afa97b8a1"} Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.372869 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-hw4dk" event={"ID":"0e51874f-4bce-4f5c-8e00-99f74b1f2881","Type":"ContainerStarted","Data":"2c6a9aefdd64f2133d4c181a86ed37fc907f8fac53099690a88d253e5964e25e"} Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.428440 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-5tt4q" event={"ID":"7737381d-608c-43f8-9cdf-a5612d28bd43","Type":"ContainerStarted","Data":"a7a055c027803c3929728207f3834fcce4d1a925291a3e9c127a413c593e4e19"} Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.428473 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b99c7c5b-z2gb5" event={"ID":"0193c2d7-ccc0-40bc-b4a9-e483e51803ce","Type":"ContainerStarted","Data":"a95018b6fc5cfff0c1923a21174d6522ce86bc222b14ea4f65f43474efca96d5"} Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.455779 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" event={"ID":"f83ce425-101d-4489-94a4-5c256eb29328","Type":"ContainerStarted","Data":"2d9c8f9e3115f67c6ef88d1a8e4688cf6f5cba6faca4d6c650065b45bf9b7b1b"} Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.498569 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-6qnjp" event={"ID":"4a8f2a74-9110-4f79-bc84-c68ed74b982d","Type":"ContainerStarted","Data":"3a69cdee6fa90a34d2a09d6dd1245ce24a1dff669016553bdd119162bd7df588"} Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.511272 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-lz5ls" event={"ID":"54002291-52de-4cd8-9c01-7a3bd3c2b98b","Type":"ContainerStarted","Data":"bd4893f94597524bc8af80091b0e92d8be69b55f72977f3a8e9f149645e529dc"} Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.530120 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6b57c44d8-px47q" event={"ID":"4b8037b3-28b8-4492-b4a7-7ed5413df620","Type":"ContainerStarted","Data":"cd0f9fc8e8e546c61297de814bdbd96e9305e429755740b4962fb53a3338f388"} Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.530159 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6b57c44d8-px47q" event={"ID":"4b8037b3-28b8-4492-b4a7-7ed5413df620","Type":"ContainerStarted","Data":"f63aa63cb97742849014f8a0b7249b2aebc9d1adbe4c36308c6bf2c1795eb660"} Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.530861 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6b57c44d8-px47q" Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.544271 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-mnftb" event={"ID":"6253eb77-9a32-473f-8f30-837d648e51cb","Type":"ContainerStarted","Data":"eaddbce93bcdbde49dce8dfc8b22483b687e91cb82b1ad8733c594cb68322d61"} Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.558531 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zs4xs" event={"ID":"98a4a627-67dd-4dac-a7a4-cdfc4baaa057","Type":"ContainerStarted","Data":"cc50038a21de7cc2d3c6e531d89382b0703488fea5a2aaefa7d1d47f971b9279"} Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.560357 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hsmt5" event={"ID":"23f6eba8-a786-4234-869f-6e44016536dd","Type":"ContainerStarted","Data":"89f2d493fcfba007a1a9c1858703e6cb7a24c4c0a19a6d6e6561043bb384943a"} Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.563048 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-dqz59" event={"ID":"43dd50ad-2f9d-47a6-b685-379dd593449c","Type":"ContainerStarted","Data":"f6f42c2cc8ef843396c2de180fcc7382ca44247176762d91833ca9d84c5fc4f9"} Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.563934 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-88c7-dqz59" Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.567136 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-kjrph" event={"ID":"e2c80909-0584-4f0c-a989-fe053b196f98","Type":"ContainerStarted","Data":"e1c1a52935c5de78f0f2afe1baa4d3d483db868040c56df1b3746b89576ea04a"} Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.580206 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9t5j7" event={"ID":"d2828482-d7e8-4e5f-985d-197d614d8d15","Type":"ContainerStarted","Data":"68b1bf70f1e556a67a88f7fd40d26b9746c54fa347d56c1a0fd212ec82619a5a"} Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.581560 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-2zhp9" event={"ID":"7546f33a-e8d7-4e15-bb6f-bd91a95d7ae9","Type":"ContainerStarted","Data":"5e57ea5d84443a99bf51716e31a2db93b71849e52dc9a8d898a1932c4894ef25"} Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.582748 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-skv6k" event={"ID":"1ec1bdf1-04ec-4b70-8e58-5df6541b021f","Type":"ContainerStarted","Data":"99116a8afc8bef5c72cc24c499ed914eddc59f2cde05f56c3e16731f1296b649"} Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.608733 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6b57c44d8-px47q" podStartSLOduration=23.608711421 podStartE2EDuration="23.608711421s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:55:03.601475656 +0000 UTC m=+870.927004656" watchObservedRunningTime="2025-10-02 16:55:03.608711421 +0000 UTC m=+870.934240421" Oct 02 16:55:03 crc kubenswrapper[4808]: I1002 16:55:03.632611 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-88c7-dqz59" podStartSLOduration=3.288863249 podStartE2EDuration="23.632591946s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:41.797709298 +0000 UTC m=+849.123238298" lastFinishedPulling="2025-10-02 16:55:02.141437965 +0000 UTC m=+869.466966995" observedRunningTime="2025-10-02 16:55:03.63017756 +0000 UTC m=+870.955706560" watchObservedRunningTime="2025-10-02 16:55:03.632591946 +0000 UTC m=+870.958120946" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.594977 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-kjrph" event={"ID":"e2c80909-0584-4f0c-a989-fe053b196f98","Type":"ContainerStarted","Data":"9819c9730c8e82849a094b4b20f8217bc028bb94d387b74411bf0c1069489ad5"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.596252 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-kjrph" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.607148 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gnv8z" event={"ID":"e1981fc0-4eab-4cec-8c2a-611ab0d2af38","Type":"ContainerStarted","Data":"b57ca7f9dc04aef8acbc8db6e54b0d0bdfd9cde475c1c42f38b5a1ab8b9f94ab"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.618461 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-5tt4q" event={"ID":"7737381d-608c-43f8-9cdf-a5612d28bd43","Type":"ContainerStarted","Data":"823a6e8daf690369e1ffcd2ac763d80c17251606e42cf7089279e28c011633d4"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.619126 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-5tt4q" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.625859 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-wpps5" event={"ID":"14931328-29f9-4a03-9a3b-50cb2bb861fe","Type":"ContainerStarted","Data":"da18e131e2bbbc38b7f8437f26cedd507c075a7286128034e24d35ecfad967c2"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.626857 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b99c7c5b-z2gb5" event={"ID":"0193c2d7-ccc0-40bc-b4a9-e483e51803ce","Type":"ContainerStarted","Data":"9c38cf50bb6ca067be3c4af0228e3c609a7a97240fd9f278e06020cf7e36d0a4"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.627381 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84b99c7c5b-z2gb5" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.633870 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" event={"ID":"818e02a3-653a-4b36-a68b-9d4e2df0b71e","Type":"ContainerStarted","Data":"bf3a5db07b00ccbb16feabeddeb0307f545e5eeeacbc3a394400c9910c2a79b3"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.634547 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.636105 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-kjrph" podStartSLOduration=4.268116494 podStartE2EDuration="24.636087375s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:41.777556544 +0000 UTC m=+849.103085544" lastFinishedPulling="2025-10-02 16:55:02.145527385 +0000 UTC m=+869.471056425" observedRunningTime="2025-10-02 16:55:04.631943263 +0000 UTC m=+871.957472263" watchObservedRunningTime="2025-10-02 16:55:04.636087375 +0000 UTC m=+871.961616375" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.650972 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-wv9sn" event={"ID":"ccfa7a5b-15b3-4fd1-9a36-f04d894a838f","Type":"ContainerStarted","Data":"7d9d268cf5d5993df41be3af27ad21b7de5987a55dcc0cbddad45884cfcb9389"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.666016 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-5tt4q" podStartSLOduration=7.145704169 podStartE2EDuration="24.666001542s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:44.795540158 +0000 UTC m=+852.121069158" lastFinishedPulling="2025-10-02 16:55:02.315837531 +0000 UTC m=+869.641366531" observedRunningTime="2025-10-02 16:55:04.664723868 +0000 UTC m=+871.990252868" watchObservedRunningTime="2025-10-02 16:55:04.666001542 +0000 UTC m=+871.991530542" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.670155 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g7gmf" event={"ID":"0cf6cc81-9904-4904-bf45-a4d3e800231b","Type":"ContainerStarted","Data":"1f556c900c9a2af8e14f9dbfbbdc0c52f6ef98c8db59f8b901137d02540f5ac3"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.677832 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-6qnjp" event={"ID":"4a8f2a74-9110-4f79-bc84-c68ed74b982d","Type":"ContainerStarted","Data":"39e9e223cb4e9b6dc1683e1105725402c3cdd631d39762130fadf28b8b23e5be"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.678465 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-6qnjp" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.681631 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-9cm78" event={"ID":"4b8b017f-fc2b-4925-ab5c-001ca052ff8c","Type":"ContainerStarted","Data":"06e267ac5c4421560c95c678535dc7eeea8054dac02968b2d0cf24ee67918542"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.682020 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-9cm78" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.683816 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vth7d" event={"ID":"8b68ec2c-6583-482b-a9f4-d75fab859c95","Type":"ContainerStarted","Data":"9c0e94872cd5450616f3eca502f772ffe595d8c32aa499dd5395de23b1919ebb"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.687449 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84b99c7c5b-z2gb5" podStartSLOduration=19.066521085 podStartE2EDuration="24.687433231s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:41.248308562 +0000 UTC m=+848.573837562" lastFinishedPulling="2025-10-02 16:54:46.869220718 +0000 UTC m=+854.194749708" observedRunningTime="2025-10-02 16:55:04.685007955 +0000 UTC m=+872.010536955" watchObservedRunningTime="2025-10-02 16:55:04.687433231 +0000 UTC m=+872.012962231" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.689957 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hsmt5" event={"ID":"23f6eba8-a786-4234-869f-6e44016536dd","Type":"ContainerStarted","Data":"c4ddf6bb3c9fb8839bc47a5d4ecadf0b5d746b6264606f8378c0e22197317074"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.690528 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hsmt5" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.705442 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-8bbm9" event={"ID":"2a264f04-66b8-4938-ab3f-58c651166227","Type":"ContainerStarted","Data":"38acfc60ca5e124e605868812ba44673ea05c014f2f82af9334f395a2dded473"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.705482 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-8bbm9" event={"ID":"2a264f04-66b8-4938-ab3f-58c651166227","Type":"ContainerStarted","Data":"e54b89413c1673ab15f187f95c0de102ce93a61e4c9bceff928ff3d1eea0271c"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.706391 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-8bbm9" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.733258 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-6qnjp" podStartSLOduration=17.540635417 podStartE2EDuration="24.733226307s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:41.408818523 +0000 UTC m=+848.734347523" lastFinishedPulling="2025-10-02 16:54:48.601409413 +0000 UTC m=+855.926938413" observedRunningTime="2025-10-02 16:55:04.732640911 +0000 UTC m=+872.058169911" watchObservedRunningTime="2025-10-02 16:55:04.733226307 +0000 UTC m=+872.058755317" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.733797 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-9cm78" podStartSLOduration=4.369807319 podStartE2EDuration="24.733791352s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:41.787868623 +0000 UTC m=+849.113397623" lastFinishedPulling="2025-10-02 16:55:02.151852646 +0000 UTC m=+869.477381656" observedRunningTime="2025-10-02 16:55:04.706618279 +0000 UTC m=+872.032147279" watchObservedRunningTime="2025-10-02 16:55:04.733791352 +0000 UTC m=+872.059320342" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.748591 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9t5j7" event={"ID":"d2828482-d7e8-4e5f-985d-197d614d8d15","Type":"ContainerStarted","Data":"f4681955a02c45e8a9d1598c0f4bbd078c4d1f393ee6afc54c9e27fd7225d965"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.748703 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9t5j7" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.751556 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-wv9sn" podStartSLOduration=5.567799257 podStartE2EDuration="24.751538371s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:43.131797559 +0000 UTC m=+850.457326549" lastFinishedPulling="2025-10-02 16:55:02.315536663 +0000 UTC m=+869.641065663" observedRunningTime="2025-10-02 16:55:04.748262882 +0000 UTC m=+872.073791892" watchObservedRunningTime="2025-10-02 16:55:04.751538371 +0000 UTC m=+872.077067371" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.764593 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-mnftb" event={"ID":"6253eb77-9a32-473f-8f30-837d648e51cb","Type":"ContainerStarted","Data":"72607dd37fa3d62a4792e7e55e1d212f02d533e73ef1978f80b29ef814d6e4ab"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.765050 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-mnftb" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.766487 4808 generic.go:334] "Generic (PLEG): container finished" podID="98a4a627-67dd-4dac-a7a4-cdfc4baaa057" containerID="c8bbe77cdb8727ad6680a1323f1fd14211b7b5a75f9768577c20bea4c40231ef" exitCode=0 Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.766911 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zs4xs" event={"ID":"98a4a627-67dd-4dac-a7a4-cdfc4baaa057","Type":"ContainerDied","Data":"c8bbe77cdb8727ad6680a1323f1fd14211b7b5a75f9768577c20bea4c40231ef"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.771089 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-2zhp9" event={"ID":"7546f33a-e8d7-4e15-bb6f-bd91a95d7ae9","Type":"ContainerStarted","Data":"4bfe7e3ee4efbf155cf9cdf9de903cf1620041c84b0b1161899060cdb6d2f0f0"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.771627 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" podStartSLOduration=4.415467481 podStartE2EDuration="24.771612742s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:41.785788877 +0000 UTC m=+849.111317877" lastFinishedPulling="2025-10-02 16:55:02.141934108 +0000 UTC m=+869.467463138" observedRunningTime="2025-10-02 16:55:04.770714608 +0000 UTC m=+872.096243608" watchObservedRunningTime="2025-10-02 16:55:04.771612742 +0000 UTC m=+872.097141742" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.771751 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-2zhp9" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.785470 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-skv6k" event={"ID":"1ec1bdf1-04ec-4b70-8e58-5df6541b021f","Type":"ContainerStarted","Data":"e49b851b184f376f273a9a90c8deafe366ea3dd9bb70f6368e5062c51bd6e7d3"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.786378 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-85777745bb-skv6k" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.789854 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-hw4dk" event={"ID":"0e51874f-4bce-4f5c-8e00-99f74b1f2881","Type":"ContainerStarted","Data":"5f535136d0554acdc0ae49cdebf89cf9a07c2c3f9b2e6e2be53ad6cd0e2cfb13"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.791816 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-hw4dk" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.816063 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-8bbm9" podStartSLOduration=7.881616928 podStartE2EDuration="24.816039231s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:45.381544232 +0000 UTC m=+852.707073232" lastFinishedPulling="2025-10-02 16:55:02.315966525 +0000 UTC m=+869.641495535" observedRunningTime="2025-10-02 16:55:04.800253145 +0000 UTC m=+872.125782145" watchObservedRunningTime="2025-10-02 16:55:04.816039231 +0000 UTC m=+872.141568231" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.832514 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-dqz59" event={"ID":"43dd50ad-2f9d-47a6-b685-379dd593449c","Type":"ContainerStarted","Data":"2575fd802f8469a055f14c51585ed4ee08c4f52c1aa03b2cc0d108ba44e9e162"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.840952 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9t5j7" podStartSLOduration=7.969489759 podStartE2EDuration="24.840933723s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:45.412869607 +0000 UTC m=+852.738398627" lastFinishedPulling="2025-10-02 16:55:02.284313581 +0000 UTC m=+869.609842591" observedRunningTime="2025-10-02 16:55:04.827179332 +0000 UTC m=+872.152708352" watchObservedRunningTime="2025-10-02 16:55:04.840933723 +0000 UTC m=+872.166462723" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.848817 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-m5tw2" event={"ID":"8771edaf-8b38-4efe-9a99-ea74be174f75","Type":"ContainerStarted","Data":"f801a15d73f1d114f0b2ae202abd2f3ce571e6a1f738409b393ad94c50ae4755"} Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.861996 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hsmt5" podStartSLOduration=4.5015926440000005 podStartE2EDuration="24.861984081s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:41.785080477 +0000 UTC m=+849.110609477" lastFinishedPulling="2025-10-02 16:55:02.145471914 +0000 UTC m=+869.471000914" observedRunningTime="2025-10-02 16:55:04.860009698 +0000 UTC m=+872.185538698" watchObservedRunningTime="2025-10-02 16:55:04.861984081 +0000 UTC m=+872.187513071" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.905274 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-mnftb" podStartSLOduration=7.970905908 podStartE2EDuration="24.905260289s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:45.381959973 +0000 UTC m=+852.707488973" lastFinishedPulling="2025-10-02 16:55:02.316314354 +0000 UTC m=+869.641843354" observedRunningTime="2025-10-02 16:55:04.903213994 +0000 UTC m=+872.228742984" watchObservedRunningTime="2025-10-02 16:55:04.905260289 +0000 UTC m=+872.230789289" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.927778 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-85777745bb-skv6k" podStartSLOduration=7.472117587 podStartE2EDuration="24.927761776s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:44.800498341 +0000 UTC m=+852.126027341" lastFinishedPulling="2025-10-02 16:55:02.25614251 +0000 UTC m=+869.581671530" observedRunningTime="2025-10-02 16:55:04.926593155 +0000 UTC m=+872.252122155" watchObservedRunningTime="2025-10-02 16:55:04.927761776 +0000 UTC m=+872.253290776" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.944227 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-2zhp9" podStartSLOduration=11.926386901 podStartE2EDuration="24.94420961s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:41.225715662 +0000 UTC m=+848.551244662" lastFinishedPulling="2025-10-02 16:54:54.243538371 +0000 UTC m=+861.569067371" observedRunningTime="2025-10-02 16:55:04.940563172 +0000 UTC m=+872.266092172" watchObservedRunningTime="2025-10-02 16:55:04.94420961 +0000 UTC m=+872.269738610" Oct 02 16:55:04 crc kubenswrapper[4808]: I1002 16:55:04.960507 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-hw4dk" podStartSLOduration=8.049492069 podStartE2EDuration="24.96049297s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:45.37926684 +0000 UTC m=+852.704795830" lastFinishedPulling="2025-10-02 16:55:02.290267721 +0000 UTC m=+869.615796731" observedRunningTime="2025-10-02 16:55:04.958993309 +0000 UTC m=+872.284522309" watchObservedRunningTime="2025-10-02 16:55:04.96049297 +0000 UTC m=+872.286021970" Oct 02 16:55:05 crc kubenswrapper[4808]: I1002 16:55:05.857329 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-wpps5" event={"ID":"14931328-29f9-4a03-9a3b-50cb2bb861fe","Type":"ContainerStarted","Data":"f45ad223d71be23cf5d6a0f5ae947a45b39c2cf40b5ce5585fefc0032e7d7844"} Oct 02 16:55:05 crc kubenswrapper[4808]: I1002 16:55:05.857559 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-wpps5" Oct 02 16:55:05 crc kubenswrapper[4808]: I1002 16:55:05.861262 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g7gmf" event={"ID":"0cf6cc81-9904-4904-bf45-a4d3e800231b","Type":"ContainerStarted","Data":"88e085e40f394abbeaccee77c3adbdeaf481ca3f7fa4f28daade4300b508695b"} Oct 02 16:55:05 crc kubenswrapper[4808]: I1002 16:55:05.861846 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g7gmf" Oct 02 16:55:05 crc kubenswrapper[4808]: I1002 16:55:05.864136 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-lz5ls" event={"ID":"54002291-52de-4cd8-9c01-7a3bd3c2b98b","Type":"ContainerStarted","Data":"ab10f5765b11d294e66a6c72412570bbc9b8c0bd345c6e1fc79b18831776c0dd"} Oct 02 16:55:05 crc kubenswrapper[4808]: I1002 16:55:05.864256 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-lz5ls" Oct 02 16:55:05 crc kubenswrapper[4808]: I1002 16:55:05.865992 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gnv8z" event={"ID":"e1981fc0-4eab-4cec-8c2a-611ab0d2af38","Type":"ContainerStarted","Data":"86c292151dada6e2aca41b9541596fecd27084131253fa3a9c956dc5cfe131c4"} Oct 02 16:55:05 crc kubenswrapper[4808]: I1002 16:55:05.866091 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gnv8z" Oct 02 16:55:05 crc kubenswrapper[4808]: I1002 16:55:05.868809 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vth7d" event={"ID":"8b68ec2c-6583-482b-a9f4-d75fab859c95","Type":"ContainerStarted","Data":"9a0e4f8b0c07983845a78323bd0cee2d73ecce695f28e96a0c786a37386e0e72"} Oct 02 16:55:05 crc kubenswrapper[4808]: I1002 16:55:05.869045 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vth7d" Oct 02 16:55:05 crc kubenswrapper[4808]: I1002 16:55:05.880375 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-m5tw2" event={"ID":"8771edaf-8b38-4efe-9a99-ea74be174f75","Type":"ContainerStarted","Data":"bcf3dc4e780e480e7a51739416ab7b6e3db90f98003635cd9bb7065056724d88"} Oct 02 16:55:05 crc kubenswrapper[4808]: I1002 16:55:05.891136 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-wpps5" podStartSLOduration=6.681347228 podStartE2EDuration="25.891115744s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:43.131521222 +0000 UTC m=+850.457050222" lastFinishedPulling="2025-10-02 16:55:02.341289738 +0000 UTC m=+869.666818738" observedRunningTime="2025-10-02 16:55:05.878104413 +0000 UTC m=+873.203633433" watchObservedRunningTime="2025-10-02 16:55:05.891115744 +0000 UTC m=+873.216644744" Oct 02 16:55:05 crc kubenswrapper[4808]: I1002 16:55:05.907909 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gnv8z" podStartSLOduration=7.98840331 podStartE2EDuration="25.907889556s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:44.370890778 +0000 UTC m=+851.696419778" lastFinishedPulling="2025-10-02 16:55:02.290377014 +0000 UTC m=+869.615906024" observedRunningTime="2025-10-02 16:55:05.901325729 +0000 UTC m=+873.226854749" watchObservedRunningTime="2025-10-02 16:55:05.907889556 +0000 UTC m=+873.233418556" Oct 02 16:55:05 crc kubenswrapper[4808]: I1002 16:55:05.921120 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vth7d" podStartSLOduration=6.735091198 podStartE2EDuration="25.921101773s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:43.13219443 +0000 UTC m=+850.457723430" lastFinishedPulling="2025-10-02 16:55:02.318205015 +0000 UTC m=+869.643734005" observedRunningTime="2025-10-02 16:55:05.918131323 +0000 UTC m=+873.243660333" watchObservedRunningTime="2025-10-02 16:55:05.921101773 +0000 UTC m=+873.246630773" Oct 02 16:55:05 crc kubenswrapper[4808]: I1002 16:55:05.939693 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g7gmf" podStartSLOduration=9.001748777 podStartE2EDuration="25.939673404s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:45.378368046 +0000 UTC m=+852.703897046" lastFinishedPulling="2025-10-02 16:55:02.316292673 +0000 UTC m=+869.641821673" observedRunningTime="2025-10-02 16:55:05.937114745 +0000 UTC m=+873.262643765" watchObservedRunningTime="2025-10-02 16:55:05.939673404 +0000 UTC m=+873.265202404" Oct 02 16:55:05 crc kubenswrapper[4808]: I1002 16:55:05.956477 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-lz5ls" podStartSLOduration=5.588355243 podStartE2EDuration="25.956460187s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:41.948844277 +0000 UTC m=+849.274373277" lastFinishedPulling="2025-10-02 16:55:02.316949221 +0000 UTC m=+869.642478221" observedRunningTime="2025-10-02 16:55:05.953350963 +0000 UTC m=+873.278879983" watchObservedRunningTime="2025-10-02 16:55:05.956460187 +0000 UTC m=+873.281989187" Oct 02 16:55:05 crc kubenswrapper[4808]: I1002 16:55:05.971061 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-m5tw2" podStartSLOduration=9.059708421 podStartE2EDuration="25.971047451s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:45.378933301 +0000 UTC m=+852.704462301" lastFinishedPulling="2025-10-02 16:55:02.290272331 +0000 UTC m=+869.615801331" observedRunningTime="2025-10-02 16:55:05.96841372 +0000 UTC m=+873.293942740" watchObservedRunningTime="2025-10-02 16:55:05.971047451 +0000 UTC m=+873.296576451" Oct 02 16:55:06 crc kubenswrapper[4808]: I1002 16:55:06.891373 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-zw4d5" event={"ID":"56b8aff7-4d44-41c4-b4a7-caed261ab974","Type":"ContainerStarted","Data":"a35571412e0ebf77fc4ec657e18eb749dad3122b7d1a90a76d1d26932dba15b8"} Oct 02 16:55:06 crc kubenswrapper[4808]: I1002 16:55:06.891868 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-zw4d5" Oct 02 16:55:06 crc kubenswrapper[4808]: I1002 16:55:06.891898 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-zw4d5" event={"ID":"56b8aff7-4d44-41c4-b4a7-caed261ab974","Type":"ContainerStarted","Data":"cc4706b7da7727782afa81f86201e2b692a8dcfb0536e46798489342919453aa"} Oct 02 16:55:06 crc kubenswrapper[4808]: I1002 16:55:06.894843 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zs4xs" event={"ID":"98a4a627-67dd-4dac-a7a4-cdfc4baaa057","Type":"ContainerStarted","Data":"bc228415fcaa92e88ec4dd0fbd801ab0665126b012fdceaafd7b1d11f134be29"} Oct 02 16:55:06 crc kubenswrapper[4808]: I1002 16:55:06.896932 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-m5tw2" Oct 02 16:55:06 crc kubenswrapper[4808]: I1002 16:55:06.940943 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-zw4d5" podStartSLOduration=15.414436469 podStartE2EDuration="26.940919324s" podCreationTimestamp="2025-10-02 16:54:40 +0000 UTC" firstStartedPulling="2025-10-02 16:54:54.272736509 +0000 UTC m=+861.598265509" lastFinishedPulling="2025-10-02 16:55:05.799219364 +0000 UTC m=+873.124748364" observedRunningTime="2025-10-02 16:55:06.936158966 +0000 UTC m=+874.261688006" watchObservedRunningTime="2025-10-02 16:55:06.940919324 +0000 UTC m=+874.266448364" Oct 02 16:55:07 crc kubenswrapper[4808]: I1002 16:55:07.903890 4808 generic.go:334] "Generic (PLEG): container finished" podID="98a4a627-67dd-4dac-a7a4-cdfc4baaa057" containerID="bc228415fcaa92e88ec4dd0fbd801ab0665126b012fdceaafd7b1d11f134be29" exitCode=0 Oct 02 16:55:07 crc kubenswrapper[4808]: I1002 16:55:07.903972 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zs4xs" event={"ID":"98a4a627-67dd-4dac-a7a4-cdfc4baaa057","Type":"ContainerDied","Data":"bc228415fcaa92e88ec4dd0fbd801ab0665126b012fdceaafd7b1d11f134be29"} Oct 02 16:55:09 crc kubenswrapper[4808]: I1002 16:55:09.923727 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zs4xs" event={"ID":"98a4a627-67dd-4dac-a7a4-cdfc4baaa057","Type":"ContainerStarted","Data":"2713f590efe7a2e35c27a4227c8dffbbf1edbb76bb2708fcbf8386cf6de35b25"} Oct 02 16:55:09 crc kubenswrapper[4808]: I1002 16:55:09.959705 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zs4xs" podStartSLOduration=23.835779386 podStartE2EDuration="26.959684098s" podCreationTimestamp="2025-10-02 16:54:43 +0000 UTC" firstStartedPulling="2025-10-02 16:55:05.706926143 +0000 UTC m=+873.032455143" lastFinishedPulling="2025-10-02 16:55:08.830830855 +0000 UTC m=+876.156359855" observedRunningTime="2025-10-02 16:55:09.95902214 +0000 UTC m=+877.284551170" watchObservedRunningTime="2025-10-02 16:55:09.959684098 +0000 UTC m=+877.285213108" Oct 02 16:55:10 crc kubenswrapper[4808]: I1002 16:55:10.567313 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84b99c7c5b-z2gb5" Oct 02 16:55:10 crc kubenswrapper[4808]: I1002 16:55:10.592579 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-2zhp9" Oct 02 16:55:10 crc kubenswrapper[4808]: I1002 16:55:10.648652 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-6qnjp" Oct 02 16:55:10 crc kubenswrapper[4808]: I1002 16:55:10.665157 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-kjrph" Oct 02 16:55:10 crc kubenswrapper[4808]: I1002 16:55:10.692940 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-9cm78" Oct 02 16:55:10 crc kubenswrapper[4808]: I1002 16:55:10.720514 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 16:55:10 crc kubenswrapper[4808]: I1002 16:55:10.737025 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-wpps5" Oct 02 16:55:10 crc kubenswrapper[4808]: I1002 16:55:10.757997 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-88c7-dqz59" Oct 02 16:55:10 crc kubenswrapper[4808]: I1002 16:55:10.769363 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-lz5ls" Oct 02 16:55:10 crc kubenswrapper[4808]: I1002 16:55:10.797503 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hsmt5" Oct 02 16:55:10 crc kubenswrapper[4808]: I1002 16:55:10.801553 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gnv8z" Oct 02 16:55:10 crc kubenswrapper[4808]: I1002 16:55:10.801604 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-mnftb" Oct 02 16:55:10 crc kubenswrapper[4808]: I1002 16:55:10.841566 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-5tt4q" Oct 02 16:55:10 crc kubenswrapper[4808]: I1002 16:55:10.863974 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-hw4dk" Oct 02 16:55:10 crc kubenswrapper[4808]: I1002 16:55:10.902673 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-m5tw2" Oct 02 16:55:10 crc kubenswrapper[4808]: I1002 16:55:10.968949 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g7gmf" Oct 02 16:55:10 crc kubenswrapper[4808]: I1002 16:55:10.976880 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-8bbm9" Oct 02 16:55:11 crc kubenswrapper[4808]: I1002 16:55:11.022505 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vth7d" Oct 02 16:55:11 crc kubenswrapper[4808]: I1002 16:55:11.061250 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-85777745bb-skv6k" Oct 02 16:55:11 crc kubenswrapper[4808]: I1002 16:55:11.135618 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9t5j7" Oct 02 16:55:11 crc kubenswrapper[4808]: I1002 16:55:11.817344 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6b57c44d8-px47q" Oct 02 16:55:12 crc kubenswrapper[4808]: I1002 16:55:12.378164 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-zw4d5" Oct 02 16:55:13 crc kubenswrapper[4808]: I1002 16:55:13.435795 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zs4xs" Oct 02 16:55:13 crc kubenswrapper[4808]: I1002 16:55:13.436323 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zs4xs" Oct 02 16:55:13 crc kubenswrapper[4808]: I1002 16:55:13.509126 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zs4xs" Oct 02 16:55:23 crc kubenswrapper[4808]: I1002 16:55:23.483876 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zs4xs" Oct 02 16:55:28 crc kubenswrapper[4808]: I1002 16:55:28.491809 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zs4xs"] Oct 02 16:55:28 crc kubenswrapper[4808]: I1002 16:55:28.492832 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zs4xs" podUID="98a4a627-67dd-4dac-a7a4-cdfc4baaa057" containerName="registry-server" containerID="cri-o://2713f590efe7a2e35c27a4227c8dffbbf1edbb76bb2708fcbf8386cf6de35b25" gracePeriod=2 Oct 02 16:55:28 crc kubenswrapper[4808]: I1002 16:55:28.920105 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zs4xs" Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.037549 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98a4a627-67dd-4dac-a7a4-cdfc4baaa057-catalog-content\") pod \"98a4a627-67dd-4dac-a7a4-cdfc4baaa057\" (UID: \"98a4a627-67dd-4dac-a7a4-cdfc4baaa057\") " Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.037646 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98a4a627-67dd-4dac-a7a4-cdfc4baaa057-utilities\") pod \"98a4a627-67dd-4dac-a7a4-cdfc4baaa057\" (UID: \"98a4a627-67dd-4dac-a7a4-cdfc4baaa057\") " Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.037750 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zx5tj\" (UniqueName: \"kubernetes.io/projected/98a4a627-67dd-4dac-a7a4-cdfc4baaa057-kube-api-access-zx5tj\") pod \"98a4a627-67dd-4dac-a7a4-cdfc4baaa057\" (UID: \"98a4a627-67dd-4dac-a7a4-cdfc4baaa057\") " Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.039312 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98a4a627-67dd-4dac-a7a4-cdfc4baaa057-utilities" (OuterVolumeSpecName: "utilities") pod "98a4a627-67dd-4dac-a7a4-cdfc4baaa057" (UID: "98a4a627-67dd-4dac-a7a4-cdfc4baaa057"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.046501 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98a4a627-67dd-4dac-a7a4-cdfc4baaa057-kube-api-access-zx5tj" (OuterVolumeSpecName: "kube-api-access-zx5tj") pod "98a4a627-67dd-4dac-a7a4-cdfc4baaa057" (UID: "98a4a627-67dd-4dac-a7a4-cdfc4baaa057"). InnerVolumeSpecName "kube-api-access-zx5tj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.091778 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98a4a627-67dd-4dac-a7a4-cdfc4baaa057-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98a4a627-67dd-4dac-a7a4-cdfc4baaa057" (UID: "98a4a627-67dd-4dac-a7a4-cdfc4baaa057"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.099945 4808 generic.go:334] "Generic (PLEG): container finished" podID="98a4a627-67dd-4dac-a7a4-cdfc4baaa057" containerID="2713f590efe7a2e35c27a4227c8dffbbf1edbb76bb2708fcbf8386cf6de35b25" exitCode=0 Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.100010 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zs4xs" event={"ID":"98a4a627-67dd-4dac-a7a4-cdfc4baaa057","Type":"ContainerDied","Data":"2713f590efe7a2e35c27a4227c8dffbbf1edbb76bb2708fcbf8386cf6de35b25"} Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.100031 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zs4xs" Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.100066 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zs4xs" event={"ID":"98a4a627-67dd-4dac-a7a4-cdfc4baaa057","Type":"ContainerDied","Data":"cc50038a21de7cc2d3c6e531d89382b0703488fea5a2aaefa7d1d47f971b9279"} Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.100097 4808 scope.go:117] "RemoveContainer" containerID="2713f590efe7a2e35c27a4227c8dffbbf1edbb76bb2708fcbf8386cf6de35b25" Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.130785 4808 scope.go:117] "RemoveContainer" containerID="bc228415fcaa92e88ec4dd0fbd801ab0665126b012fdceaafd7b1d11f134be29" Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.140470 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98a4a627-67dd-4dac-a7a4-cdfc4baaa057-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.140531 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zx5tj\" (UniqueName: \"kubernetes.io/projected/98a4a627-67dd-4dac-a7a4-cdfc4baaa057-kube-api-access-zx5tj\") on node \"crc\" DevicePath \"\"" Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.140546 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98a4a627-67dd-4dac-a7a4-cdfc4baaa057-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.149381 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zs4xs"] Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.153261 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zs4xs"] Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.170303 4808 scope.go:117] "RemoveContainer" containerID="c8bbe77cdb8727ad6680a1323f1fd14211b7b5a75f9768577c20bea4c40231ef" Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.191477 4808 scope.go:117] "RemoveContainer" containerID="2713f590efe7a2e35c27a4227c8dffbbf1edbb76bb2708fcbf8386cf6de35b25" Oct 02 16:55:29 crc kubenswrapper[4808]: E1002 16:55:29.191813 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2713f590efe7a2e35c27a4227c8dffbbf1edbb76bb2708fcbf8386cf6de35b25\": container with ID starting with 2713f590efe7a2e35c27a4227c8dffbbf1edbb76bb2708fcbf8386cf6de35b25 not found: ID does not exist" containerID="2713f590efe7a2e35c27a4227c8dffbbf1edbb76bb2708fcbf8386cf6de35b25" Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.191853 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2713f590efe7a2e35c27a4227c8dffbbf1edbb76bb2708fcbf8386cf6de35b25"} err="failed to get container status \"2713f590efe7a2e35c27a4227c8dffbbf1edbb76bb2708fcbf8386cf6de35b25\": rpc error: code = NotFound desc = could not find container \"2713f590efe7a2e35c27a4227c8dffbbf1edbb76bb2708fcbf8386cf6de35b25\": container with ID starting with 2713f590efe7a2e35c27a4227c8dffbbf1edbb76bb2708fcbf8386cf6de35b25 not found: ID does not exist" Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.191882 4808 scope.go:117] "RemoveContainer" containerID="bc228415fcaa92e88ec4dd0fbd801ab0665126b012fdceaafd7b1d11f134be29" Oct 02 16:55:29 crc kubenswrapper[4808]: E1002 16:55:29.192201 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc228415fcaa92e88ec4dd0fbd801ab0665126b012fdceaafd7b1d11f134be29\": container with ID starting with bc228415fcaa92e88ec4dd0fbd801ab0665126b012fdceaafd7b1d11f134be29 not found: ID does not exist" containerID="bc228415fcaa92e88ec4dd0fbd801ab0665126b012fdceaafd7b1d11f134be29" Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.192256 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc228415fcaa92e88ec4dd0fbd801ab0665126b012fdceaafd7b1d11f134be29"} err="failed to get container status \"bc228415fcaa92e88ec4dd0fbd801ab0665126b012fdceaafd7b1d11f134be29\": rpc error: code = NotFound desc = could not find container \"bc228415fcaa92e88ec4dd0fbd801ab0665126b012fdceaafd7b1d11f134be29\": container with ID starting with bc228415fcaa92e88ec4dd0fbd801ab0665126b012fdceaafd7b1d11f134be29 not found: ID does not exist" Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.192288 4808 scope.go:117] "RemoveContainer" containerID="c8bbe77cdb8727ad6680a1323f1fd14211b7b5a75f9768577c20bea4c40231ef" Oct 02 16:55:29 crc kubenswrapper[4808]: E1002 16:55:29.192583 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8bbe77cdb8727ad6680a1323f1fd14211b7b5a75f9768577c20bea4c40231ef\": container with ID starting with c8bbe77cdb8727ad6680a1323f1fd14211b7b5a75f9768577c20bea4c40231ef not found: ID does not exist" containerID="c8bbe77cdb8727ad6680a1323f1fd14211b7b5a75f9768577c20bea4c40231ef" Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.192631 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8bbe77cdb8727ad6680a1323f1fd14211b7b5a75f9768577c20bea4c40231ef"} err="failed to get container status \"c8bbe77cdb8727ad6680a1323f1fd14211b7b5a75f9768577c20bea4c40231ef\": rpc error: code = NotFound desc = could not find container \"c8bbe77cdb8727ad6680a1323f1fd14211b7b5a75f9768577c20bea4c40231ef\": container with ID starting with c8bbe77cdb8727ad6680a1323f1fd14211b7b5a75f9768577c20bea4c40231ef not found: ID does not exist" Oct 02 16:55:29 crc kubenswrapper[4808]: I1002 16:55:29.408516 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98a4a627-67dd-4dac-a7a4-cdfc4baaa057" path="/var/lib/kubelet/pods/98a4a627-67dd-4dac-a7a4-cdfc4baaa057/volumes" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.662697 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-x6bs8"] Oct 02 16:55:30 crc kubenswrapper[4808]: E1002 16:55:30.663019 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98a4a627-67dd-4dac-a7a4-cdfc4baaa057" containerName="registry-server" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.663034 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="98a4a627-67dd-4dac-a7a4-cdfc4baaa057" containerName="registry-server" Oct 02 16:55:30 crc kubenswrapper[4808]: E1002 16:55:30.663089 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98a4a627-67dd-4dac-a7a4-cdfc4baaa057" containerName="extract-content" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.663097 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="98a4a627-67dd-4dac-a7a4-cdfc4baaa057" containerName="extract-content" Oct 02 16:55:30 crc kubenswrapper[4808]: E1002 16:55:30.663116 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98a4a627-67dd-4dac-a7a4-cdfc4baaa057" containerName="extract-utilities" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.663124 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="98a4a627-67dd-4dac-a7a4-cdfc4baaa057" containerName="extract-utilities" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.663355 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="98a4a627-67dd-4dac-a7a4-cdfc4baaa057" containerName="registry-server" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.664192 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-x6bs8" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.666434 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.666472 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.667405 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.667603 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-82xgn" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.674545 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-x6bs8"] Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.724529 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-hkgw6"] Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.725569 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-hkgw6" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.727607 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.736269 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-hkgw6"] Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.762037 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qjx8\" (UniqueName: \"kubernetes.io/projected/1636ee11-eba2-4ab3-9546-d824dbc5593b-kube-api-access-4qjx8\") pod \"dnsmasq-dns-675f4bcbfc-x6bs8\" (UID: \"1636ee11-eba2-4ab3-9546-d824dbc5593b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-x6bs8" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.762119 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1636ee11-eba2-4ab3-9546-d824dbc5593b-config\") pod \"dnsmasq-dns-675f4bcbfc-x6bs8\" (UID: \"1636ee11-eba2-4ab3-9546-d824dbc5593b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-x6bs8" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.862853 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e85a2bd7-ef57-4edd-b643-0def775bd59b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-hkgw6\" (UID: \"e85a2bd7-ef57-4edd-b643-0def775bd59b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hkgw6" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.862901 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wwk4\" (UniqueName: \"kubernetes.io/projected/e85a2bd7-ef57-4edd-b643-0def775bd59b-kube-api-access-6wwk4\") pod \"dnsmasq-dns-78dd6ddcc-hkgw6\" (UID: \"e85a2bd7-ef57-4edd-b643-0def775bd59b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hkgw6" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.862940 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1636ee11-eba2-4ab3-9546-d824dbc5593b-config\") pod \"dnsmasq-dns-675f4bcbfc-x6bs8\" (UID: \"1636ee11-eba2-4ab3-9546-d824dbc5593b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-x6bs8" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.862991 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e85a2bd7-ef57-4edd-b643-0def775bd59b-config\") pod \"dnsmasq-dns-78dd6ddcc-hkgw6\" (UID: \"e85a2bd7-ef57-4edd-b643-0def775bd59b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hkgw6" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.863026 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qjx8\" (UniqueName: \"kubernetes.io/projected/1636ee11-eba2-4ab3-9546-d824dbc5593b-kube-api-access-4qjx8\") pod \"dnsmasq-dns-675f4bcbfc-x6bs8\" (UID: \"1636ee11-eba2-4ab3-9546-d824dbc5593b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-x6bs8" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.863752 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1636ee11-eba2-4ab3-9546-d824dbc5593b-config\") pod \"dnsmasq-dns-675f4bcbfc-x6bs8\" (UID: \"1636ee11-eba2-4ab3-9546-d824dbc5593b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-x6bs8" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.881435 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qjx8\" (UniqueName: \"kubernetes.io/projected/1636ee11-eba2-4ab3-9546-d824dbc5593b-kube-api-access-4qjx8\") pod \"dnsmasq-dns-675f4bcbfc-x6bs8\" (UID: \"1636ee11-eba2-4ab3-9546-d824dbc5593b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-x6bs8" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.964428 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e85a2bd7-ef57-4edd-b643-0def775bd59b-config\") pod \"dnsmasq-dns-78dd6ddcc-hkgw6\" (UID: \"e85a2bd7-ef57-4edd-b643-0def775bd59b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hkgw6" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.964514 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e85a2bd7-ef57-4edd-b643-0def775bd59b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-hkgw6\" (UID: \"e85a2bd7-ef57-4edd-b643-0def775bd59b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hkgw6" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.964537 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wwk4\" (UniqueName: \"kubernetes.io/projected/e85a2bd7-ef57-4edd-b643-0def775bd59b-kube-api-access-6wwk4\") pod \"dnsmasq-dns-78dd6ddcc-hkgw6\" (UID: \"e85a2bd7-ef57-4edd-b643-0def775bd59b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hkgw6" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.965340 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e85a2bd7-ef57-4edd-b643-0def775bd59b-config\") pod \"dnsmasq-dns-78dd6ddcc-hkgw6\" (UID: \"e85a2bd7-ef57-4edd-b643-0def775bd59b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hkgw6" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.965960 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e85a2bd7-ef57-4edd-b643-0def775bd59b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-hkgw6\" (UID: \"e85a2bd7-ef57-4edd-b643-0def775bd59b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hkgw6" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.982096 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wwk4\" (UniqueName: \"kubernetes.io/projected/e85a2bd7-ef57-4edd-b643-0def775bd59b-kube-api-access-6wwk4\") pod \"dnsmasq-dns-78dd6ddcc-hkgw6\" (UID: \"e85a2bd7-ef57-4edd-b643-0def775bd59b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hkgw6" Oct 02 16:55:30 crc kubenswrapper[4808]: I1002 16:55:30.982323 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-x6bs8" Oct 02 16:55:31 crc kubenswrapper[4808]: I1002 16:55:31.048881 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-hkgw6" Oct 02 16:55:31 crc kubenswrapper[4808]: I1002 16:55:31.315946 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-hkgw6"] Oct 02 16:55:31 crc kubenswrapper[4808]: I1002 16:55:31.342743 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 16:55:31 crc kubenswrapper[4808]: I1002 16:55:31.505804 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-x6bs8"] Oct 02 16:55:31 crc kubenswrapper[4808]: W1002 16:55:31.512852 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1636ee11_eba2_4ab3_9546_d824dbc5593b.slice/crio-57c16fd7f7cded477ce0ce4523173eeea7ebd27bb4b88edac5327e0009863fc4 WatchSource:0}: Error finding container 57c16fd7f7cded477ce0ce4523173eeea7ebd27bb4b88edac5327e0009863fc4: Status 404 returned error can't find the container with id 57c16fd7f7cded477ce0ce4523173eeea7ebd27bb4b88edac5327e0009863fc4 Oct 02 16:55:32 crc kubenswrapper[4808]: I1002 16:55:32.128431 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-hkgw6" event={"ID":"e85a2bd7-ef57-4edd-b643-0def775bd59b","Type":"ContainerStarted","Data":"bb6742ce0ca1834d0e8ee6737fbf8fe679728dbd6d16547be4689afc07c4595d"} Oct 02 16:55:32 crc kubenswrapper[4808]: I1002 16:55:32.130412 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-x6bs8" event={"ID":"1636ee11-eba2-4ab3-9546-d824dbc5593b","Type":"ContainerStarted","Data":"57c16fd7f7cded477ce0ce4523173eeea7ebd27bb4b88edac5327e0009863fc4"} Oct 02 16:55:33 crc kubenswrapper[4808]: I1002 16:55:33.636540 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-x6bs8"] Oct 02 16:55:33 crc kubenswrapper[4808]: I1002 16:55:33.665416 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-t2nm2"] Oct 02 16:55:33 crc kubenswrapper[4808]: I1002 16:55:33.666639 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" Oct 02 16:55:33 crc kubenswrapper[4808]: I1002 16:55:33.671150 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-t2nm2"] Oct 02 16:55:33 crc kubenswrapper[4808]: I1002 16:55:33.836633 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmn2m\" (UniqueName: \"kubernetes.io/projected/b11c2afc-30b4-4795-870c-1b710a87c054-kube-api-access-bmn2m\") pod \"dnsmasq-dns-666b6646f7-t2nm2\" (UID: \"b11c2afc-30b4-4795-870c-1b710a87c054\") " pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" Oct 02 16:55:33 crc kubenswrapper[4808]: I1002 16:55:33.836687 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b11c2afc-30b4-4795-870c-1b710a87c054-config\") pod \"dnsmasq-dns-666b6646f7-t2nm2\" (UID: \"b11c2afc-30b4-4795-870c-1b710a87c054\") " pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" Oct 02 16:55:33 crc kubenswrapper[4808]: I1002 16:55:33.836715 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b11c2afc-30b4-4795-870c-1b710a87c054-dns-svc\") pod \"dnsmasq-dns-666b6646f7-t2nm2\" (UID: \"b11c2afc-30b4-4795-870c-1b710a87c054\") " pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" Oct 02 16:55:33 crc kubenswrapper[4808]: I1002 16:55:33.937929 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmn2m\" (UniqueName: \"kubernetes.io/projected/b11c2afc-30b4-4795-870c-1b710a87c054-kube-api-access-bmn2m\") pod \"dnsmasq-dns-666b6646f7-t2nm2\" (UID: \"b11c2afc-30b4-4795-870c-1b710a87c054\") " pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" Oct 02 16:55:33 crc kubenswrapper[4808]: I1002 16:55:33.937976 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b11c2afc-30b4-4795-870c-1b710a87c054-config\") pod \"dnsmasq-dns-666b6646f7-t2nm2\" (UID: \"b11c2afc-30b4-4795-870c-1b710a87c054\") " pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" Oct 02 16:55:33 crc kubenswrapper[4808]: I1002 16:55:33.940366 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b11c2afc-30b4-4795-870c-1b710a87c054-dns-svc\") pod \"dnsmasq-dns-666b6646f7-t2nm2\" (UID: \"b11c2afc-30b4-4795-870c-1b710a87c054\") " pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" Oct 02 16:55:33 crc kubenswrapper[4808]: I1002 16:55:33.940921 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b11c2afc-30b4-4795-870c-1b710a87c054-config\") pod \"dnsmasq-dns-666b6646f7-t2nm2\" (UID: \"b11c2afc-30b4-4795-870c-1b710a87c054\") " pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" Oct 02 16:55:33 crc kubenswrapper[4808]: I1002 16:55:33.941126 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b11c2afc-30b4-4795-870c-1b710a87c054-dns-svc\") pod \"dnsmasq-dns-666b6646f7-t2nm2\" (UID: \"b11c2afc-30b4-4795-870c-1b710a87c054\") " pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" Oct 02 16:55:33 crc kubenswrapper[4808]: I1002 16:55:33.961831 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmn2m\" (UniqueName: \"kubernetes.io/projected/b11c2afc-30b4-4795-870c-1b710a87c054-kube-api-access-bmn2m\") pod \"dnsmasq-dns-666b6646f7-t2nm2\" (UID: \"b11c2afc-30b4-4795-870c-1b710a87c054\") " pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" Oct 02 16:55:33 crc kubenswrapper[4808]: I1002 16:55:33.961930 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-hkgw6"] Oct 02 16:55:33 crc kubenswrapper[4808]: I1002 16:55:33.989016 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-xgpxx"] Oct 02 16:55:33 crc kubenswrapper[4808]: I1002 16:55:33.992471 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" Oct 02 16:55:33 crc kubenswrapper[4808]: I1002 16:55:33.999727 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.007589 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-xgpxx"] Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.041092 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brfn7\" (UniqueName: \"kubernetes.io/projected/45c90d4a-2e61-431a-97b4-900c0ceb6493-kube-api-access-brfn7\") pod \"dnsmasq-dns-57d769cc4f-xgpxx\" (UID: \"45c90d4a-2e61-431a-97b4-900c0ceb6493\") " pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.041435 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45c90d4a-2e61-431a-97b4-900c0ceb6493-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-xgpxx\" (UID: \"45c90d4a-2e61-431a-97b4-900c0ceb6493\") " pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.041476 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45c90d4a-2e61-431a-97b4-900c0ceb6493-config\") pod \"dnsmasq-dns-57d769cc4f-xgpxx\" (UID: \"45c90d4a-2e61-431a-97b4-900c0ceb6493\") " pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.142506 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45c90d4a-2e61-431a-97b4-900c0ceb6493-config\") pod \"dnsmasq-dns-57d769cc4f-xgpxx\" (UID: \"45c90d4a-2e61-431a-97b4-900c0ceb6493\") " pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.142603 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brfn7\" (UniqueName: \"kubernetes.io/projected/45c90d4a-2e61-431a-97b4-900c0ceb6493-kube-api-access-brfn7\") pod \"dnsmasq-dns-57d769cc4f-xgpxx\" (UID: \"45c90d4a-2e61-431a-97b4-900c0ceb6493\") " pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.142634 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45c90d4a-2e61-431a-97b4-900c0ceb6493-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-xgpxx\" (UID: \"45c90d4a-2e61-431a-97b4-900c0ceb6493\") " pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.143376 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45c90d4a-2e61-431a-97b4-900c0ceb6493-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-xgpxx\" (UID: \"45c90d4a-2e61-431a-97b4-900c0ceb6493\") " pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.143922 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45c90d4a-2e61-431a-97b4-900c0ceb6493-config\") pod \"dnsmasq-dns-57d769cc4f-xgpxx\" (UID: \"45c90d4a-2e61-431a-97b4-900c0ceb6493\") " pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.170972 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brfn7\" (UniqueName: \"kubernetes.io/projected/45c90d4a-2e61-431a-97b4-900c0ceb6493-kube-api-access-brfn7\") pod \"dnsmasq-dns-57d769cc4f-xgpxx\" (UID: \"45c90d4a-2e61-431a-97b4-900c0ceb6493\") " pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.317640 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.538682 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-t2nm2"] Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.803851 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.805264 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.807594 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.807695 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.808000 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.810850 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.811072 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.812136 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-xzjch" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.813347 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.819747 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.955011 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.955058 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f71818d6-1ebf-4838-b871-8873c107b325-config-data\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.955113 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.955137 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f71818d6-1ebf-4838-b871-8873c107b325-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.955192 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.955290 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f71818d6-1ebf-4838-b871-8873c107b325-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.955453 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.955538 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tvxc\" (UniqueName: \"kubernetes.io/projected/f71818d6-1ebf-4838-b871-8873c107b325-kube-api-access-5tvxc\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.955569 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.955591 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f71818d6-1ebf-4838-b871-8873c107b325-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:34 crc kubenswrapper[4808]: I1002 16:55:34.955636 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f71818d6-1ebf-4838-b871-8873c107b325-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.056352 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f71818d6-1ebf-4838-b871-8873c107b325-config-data\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.056396 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.056420 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f71818d6-1ebf-4838-b871-8873c107b325-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.056436 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.056451 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f71818d6-1ebf-4838-b871-8873c107b325-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.056508 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.056545 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tvxc\" (UniqueName: \"kubernetes.io/projected/f71818d6-1ebf-4838-b871-8873c107b325-kube-api-access-5tvxc\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.056567 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.056584 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f71818d6-1ebf-4838-b871-8873c107b325-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.056612 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f71818d6-1ebf-4838-b871-8873c107b325-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.056636 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.056899 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.057087 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.058478 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.058951 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f71818d6-1ebf-4838-b871-8873c107b325-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.059416 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f71818d6-1ebf-4838-b871-8873c107b325-config-data\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.061837 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f71818d6-1ebf-4838-b871-8873c107b325-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.063938 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f71818d6-1ebf-4838-b871-8873c107b325-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.067683 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f71818d6-1ebf-4838-b871-8873c107b325-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.069008 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.070677 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.079379 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tvxc\" (UniqueName: \"kubernetes.io/projected/f71818d6-1ebf-4838-b871-8873c107b325-kube-api-access-5tvxc\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.091736 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.106826 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.109297 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.113939 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.114146 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.114415 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.114587 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.114647 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-7qq44" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.114701 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.118800 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.126127 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.134515 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.259165 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.259219 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.259259 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.259275 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.259297 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.259337 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.259411 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.259526 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.259550 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.259593 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8669v\" (UniqueName: \"kubernetes.io/projected/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-kube-api-access-8669v\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.259634 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.360549 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.360602 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.360624 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.360653 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.360707 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.360730 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.360763 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.360785 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.360806 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8669v\" (UniqueName: \"kubernetes.io/projected/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-kube-api-access-8669v\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.360836 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.360888 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.361027 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.361803 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.361926 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.362188 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.362670 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.362734 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.367175 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.367367 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.369705 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.377423 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.378172 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8669v\" (UniqueName: \"kubernetes.io/projected/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-kube-api-access-8669v\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.386403 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:35 crc kubenswrapper[4808]: I1002 16:55:35.435307 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.802993 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.804743 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.813617 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.842413 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.842531 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.847142 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.847510 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-z6kmm" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.847640 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.849626 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.899727 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw2pp\" (UniqueName: \"kubernetes.io/projected/500b4d21-bf0e-4763-91bd-7c32bf13a064-kube-api-access-hw2pp\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.899869 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/500b4d21-bf0e-4763-91bd-7c32bf13a064-config-data-generated\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.899943 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/500b4d21-bf0e-4763-91bd-7c32bf13a064-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.900014 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/500b4d21-bf0e-4763-91bd-7c32bf13a064-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.900095 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/500b4d21-bf0e-4763-91bd-7c32bf13a064-kolla-config\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.900177 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/500b4d21-bf0e-4763-91bd-7c32bf13a064-secrets\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.900206 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/500b4d21-bf0e-4763-91bd-7c32bf13a064-operator-scripts\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.900281 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/500b4d21-bf0e-4763-91bd-7c32bf13a064-config-data-default\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.900436 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.925315 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.926771 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.930529 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.931749 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.931973 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-j8zpv" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.931894 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 02 16:55:37 crc kubenswrapper[4808]: I1002 16:55:37.932480 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.001679 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/500b4d21-bf0e-4763-91bd-7c32bf13a064-config-data-generated\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.001731 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/500b4d21-bf0e-4763-91bd-7c32bf13a064-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.001763 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/500b4d21-bf0e-4763-91bd-7c32bf13a064-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.001804 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/500b4d21-bf0e-4763-91bd-7c32bf13a064-kolla-config\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.001851 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/500b4d21-bf0e-4763-91bd-7c32bf13a064-secrets\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.001878 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/500b4d21-bf0e-4763-91bd-7c32bf13a064-operator-scripts\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.001903 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a8509820-f0d7-417b-92d4-1be8b7b714ef-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.001924 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/500b4d21-bf0e-4763-91bd-7c32bf13a064-config-data-default\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.001960 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbgkg\" (UniqueName: \"kubernetes.io/projected/a8509820-f0d7-417b-92d4-1be8b7b714ef-kube-api-access-lbgkg\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.002033 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a8509820-f0d7-417b-92d4-1be8b7b714ef-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.002080 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a8509820-f0d7-417b-92d4-1be8b7b714ef-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.002119 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/a8509820-f0d7-417b-92d4-1be8b7b714ef-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.002152 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.002188 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8509820-f0d7-417b-92d4-1be8b7b714ef-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.002229 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.002299 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw2pp\" (UniqueName: \"kubernetes.io/projected/500b4d21-bf0e-4763-91bd-7c32bf13a064-kube-api-access-hw2pp\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.002333 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8509820-f0d7-417b-92d4-1be8b7b714ef-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.002333 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/500b4d21-bf0e-4763-91bd-7c32bf13a064-config-data-generated\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.002368 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8509820-f0d7-417b-92d4-1be8b7b714ef-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.003082 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.003441 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/500b4d21-bf0e-4763-91bd-7c32bf13a064-config-data-default\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.003567 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/500b4d21-bf0e-4763-91bd-7c32bf13a064-kolla-config\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.003654 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/500b4d21-bf0e-4763-91bd-7c32bf13a064-operator-scripts\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.006705 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/500b4d21-bf0e-4763-91bd-7c32bf13a064-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.011855 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/500b4d21-bf0e-4763-91bd-7c32bf13a064-secrets\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.019967 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/500b4d21-bf0e-4763-91bd-7c32bf13a064-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.021709 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.032902 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw2pp\" (UniqueName: \"kubernetes.io/projected/500b4d21-bf0e-4763-91bd-7c32bf13a064-kube-api-access-hw2pp\") pod \"openstack-galera-0\" (UID: \"500b4d21-bf0e-4763-91bd-7c32bf13a064\") " pod="openstack/openstack-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.104365 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.104437 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8509820-f0d7-417b-92d4-1be8b7b714ef-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.104462 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8509820-f0d7-417b-92d4-1be8b7b714ef-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.104517 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a8509820-f0d7-417b-92d4-1be8b7b714ef-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.104547 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbgkg\" (UniqueName: \"kubernetes.io/projected/a8509820-f0d7-417b-92d4-1be8b7b714ef-kube-api-access-lbgkg\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.104629 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a8509820-f0d7-417b-92d4-1be8b7b714ef-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.104661 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a8509820-f0d7-417b-92d4-1be8b7b714ef-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.104684 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/a8509820-f0d7-417b-92d4-1be8b7b714ef-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.104709 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8509820-f0d7-417b-92d4-1be8b7b714ef-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.107929 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a8509820-f0d7-417b-92d4-1be8b7b714ef-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.109688 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a8509820-f0d7-417b-92d4-1be8b7b714ef-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.111697 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8509820-f0d7-417b-92d4-1be8b7b714ef-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.111894 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.112226 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8509820-f0d7-417b-92d4-1be8b7b714ef-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.112371 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a8509820-f0d7-417b-92d4-1be8b7b714ef-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.119166 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8509820-f0d7-417b-92d4-1be8b7b714ef-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.120307 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/a8509820-f0d7-417b-92d4-1be8b7b714ef-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.127899 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbgkg\" (UniqueName: \"kubernetes.io/projected/a8509820-f0d7-417b-92d4-1be8b7b714ef-kube-api-access-lbgkg\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.154183 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a8509820-f0d7-417b-92d4-1be8b7b714ef\") " pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.158500 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.181661 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" event={"ID":"b11c2afc-30b4-4795-870c-1b710a87c054","Type":"ContainerStarted","Data":"433ff68fa06af4af4b1df7ad203c2ba462f930731ae2a4faa5de8a0ef88e1a5a"} Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.241258 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.317846 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.318812 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.320658 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.320863 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-zd8zs" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.321816 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.328368 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.407631 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b0cce8a-5f59-490d-87e6-6186890cd2a8-config-data\") pod \"memcached-0\" (UID: \"8b0cce8a-5f59-490d-87e6-6186890cd2a8\") " pod="openstack/memcached-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.407701 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b0cce8a-5f59-490d-87e6-6186890cd2a8-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8b0cce8a-5f59-490d-87e6-6186890cd2a8\") " pod="openstack/memcached-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.407908 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjwts\" (UniqueName: \"kubernetes.io/projected/8b0cce8a-5f59-490d-87e6-6186890cd2a8-kube-api-access-qjwts\") pod \"memcached-0\" (UID: \"8b0cce8a-5f59-490d-87e6-6186890cd2a8\") " pod="openstack/memcached-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.407964 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b0cce8a-5f59-490d-87e6-6186890cd2a8-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8b0cce8a-5f59-490d-87e6-6186890cd2a8\") " pod="openstack/memcached-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.408009 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8b0cce8a-5f59-490d-87e6-6186890cd2a8-kolla-config\") pod \"memcached-0\" (UID: \"8b0cce8a-5f59-490d-87e6-6186890cd2a8\") " pod="openstack/memcached-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.509615 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjwts\" (UniqueName: \"kubernetes.io/projected/8b0cce8a-5f59-490d-87e6-6186890cd2a8-kube-api-access-qjwts\") pod \"memcached-0\" (UID: \"8b0cce8a-5f59-490d-87e6-6186890cd2a8\") " pod="openstack/memcached-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.509664 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b0cce8a-5f59-490d-87e6-6186890cd2a8-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8b0cce8a-5f59-490d-87e6-6186890cd2a8\") " pod="openstack/memcached-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.509691 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8b0cce8a-5f59-490d-87e6-6186890cd2a8-kolla-config\") pod \"memcached-0\" (UID: \"8b0cce8a-5f59-490d-87e6-6186890cd2a8\") " pod="openstack/memcached-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.509722 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b0cce8a-5f59-490d-87e6-6186890cd2a8-config-data\") pod \"memcached-0\" (UID: \"8b0cce8a-5f59-490d-87e6-6186890cd2a8\") " pod="openstack/memcached-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.509756 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b0cce8a-5f59-490d-87e6-6186890cd2a8-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8b0cce8a-5f59-490d-87e6-6186890cd2a8\") " pod="openstack/memcached-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.510959 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b0cce8a-5f59-490d-87e6-6186890cd2a8-config-data\") pod \"memcached-0\" (UID: \"8b0cce8a-5f59-490d-87e6-6186890cd2a8\") " pod="openstack/memcached-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.511024 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8b0cce8a-5f59-490d-87e6-6186890cd2a8-kolla-config\") pod \"memcached-0\" (UID: \"8b0cce8a-5f59-490d-87e6-6186890cd2a8\") " pod="openstack/memcached-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.514184 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b0cce8a-5f59-490d-87e6-6186890cd2a8-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8b0cce8a-5f59-490d-87e6-6186890cd2a8\") " pod="openstack/memcached-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.515822 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b0cce8a-5f59-490d-87e6-6186890cd2a8-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8b0cce8a-5f59-490d-87e6-6186890cd2a8\") " pod="openstack/memcached-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.528365 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjwts\" (UniqueName: \"kubernetes.io/projected/8b0cce8a-5f59-490d-87e6-6186890cd2a8-kube-api-access-qjwts\") pod \"memcached-0\" (UID: \"8b0cce8a-5f59-490d-87e6-6186890cd2a8\") " pod="openstack/memcached-0" Oct 02 16:55:38 crc kubenswrapper[4808]: I1002 16:55:38.645570 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 16:55:40 crc kubenswrapper[4808]: I1002 16:55:40.126195 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 16:55:40 crc kubenswrapper[4808]: I1002 16:55:40.127442 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 16:55:40 crc kubenswrapper[4808]: I1002 16:55:40.130672 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-qtw8c" Oct 02 16:55:40 crc kubenswrapper[4808]: I1002 16:55:40.131124 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 16:55:40 crc kubenswrapper[4808]: I1002 16:55:40.259598 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr2wm\" (UniqueName: \"kubernetes.io/projected/f18f563e-79e6-4c33-87a5-91f7c27bf181-kube-api-access-lr2wm\") pod \"kube-state-metrics-0\" (UID: \"f18f563e-79e6-4c33-87a5-91f7c27bf181\") " pod="openstack/kube-state-metrics-0" Oct 02 16:55:40 crc kubenswrapper[4808]: I1002 16:55:40.361618 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr2wm\" (UniqueName: \"kubernetes.io/projected/f18f563e-79e6-4c33-87a5-91f7c27bf181-kube-api-access-lr2wm\") pod \"kube-state-metrics-0\" (UID: \"f18f563e-79e6-4c33-87a5-91f7c27bf181\") " pod="openstack/kube-state-metrics-0" Oct 02 16:55:40 crc kubenswrapper[4808]: I1002 16:55:40.394685 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr2wm\" (UniqueName: \"kubernetes.io/projected/f18f563e-79e6-4c33-87a5-91f7c27bf181-kube-api-access-lr2wm\") pod \"kube-state-metrics-0\" (UID: \"f18f563e-79e6-4c33-87a5-91f7c27bf181\") " pod="openstack/kube-state-metrics-0" Oct 02 16:55:40 crc kubenswrapper[4808]: I1002 16:55:40.471433 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.526450 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-hzbv4"] Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.528799 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.531589 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.531833 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.531970 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-v64tp" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.537698 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-59c8c"] Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.539520 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.542462 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-hzbv4"] Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.575874 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-59c8c"] Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.639482 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-var-run\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.639530 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk8vf\" (UniqueName: \"kubernetes.io/projected/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-kube-api-access-fk8vf\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.639559 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-combined-ca-bundle\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.639588 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd2a1d38-1169-4ab7-995f-eb6fb2f230a6-scripts\") pod \"ovn-controller-ovs-59c8c\" (UID: \"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6\") " pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.639777 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fd2a1d38-1169-4ab7-995f-eb6fb2f230a6-var-log\") pod \"ovn-controller-ovs-59c8c\" (UID: \"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6\") " pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.639815 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-var-log-ovn\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.639846 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-ovn-controller-tls-certs\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.639974 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-var-run-ovn\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.639996 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fd2a1d38-1169-4ab7-995f-eb6fb2f230a6-var-lib\") pod \"ovn-controller-ovs-59c8c\" (UID: \"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6\") " pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.640032 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fd2a1d38-1169-4ab7-995f-eb6fb2f230a6-etc-ovs\") pod \"ovn-controller-ovs-59c8c\" (UID: \"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6\") " pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.640052 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-scripts\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.640084 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6s7j\" (UniqueName: \"kubernetes.io/projected/fd2a1d38-1169-4ab7-995f-eb6fb2f230a6-kube-api-access-n6s7j\") pod \"ovn-controller-ovs-59c8c\" (UID: \"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6\") " pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.640120 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fd2a1d38-1169-4ab7-995f-eb6fb2f230a6-var-run\") pod \"ovn-controller-ovs-59c8c\" (UID: \"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6\") " pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.741927 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6s7j\" (UniqueName: \"kubernetes.io/projected/fd2a1d38-1169-4ab7-995f-eb6fb2f230a6-kube-api-access-n6s7j\") pod \"ovn-controller-ovs-59c8c\" (UID: \"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6\") " pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.741965 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fd2a1d38-1169-4ab7-995f-eb6fb2f230a6-var-run\") pod \"ovn-controller-ovs-59c8c\" (UID: \"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6\") " pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.741995 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-var-run\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.742012 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk8vf\" (UniqueName: \"kubernetes.io/projected/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-kube-api-access-fk8vf\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.742037 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-combined-ca-bundle\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.742062 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd2a1d38-1169-4ab7-995f-eb6fb2f230a6-scripts\") pod \"ovn-controller-ovs-59c8c\" (UID: \"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6\") " pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.742099 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fd2a1d38-1169-4ab7-995f-eb6fb2f230a6-var-log\") pod \"ovn-controller-ovs-59c8c\" (UID: \"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6\") " pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.742117 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-var-log-ovn\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.742133 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-ovn-controller-tls-certs\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.742175 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-var-run-ovn\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.742188 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fd2a1d38-1169-4ab7-995f-eb6fb2f230a6-var-lib\") pod \"ovn-controller-ovs-59c8c\" (UID: \"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6\") " pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.742210 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fd2a1d38-1169-4ab7-995f-eb6fb2f230a6-etc-ovs\") pod \"ovn-controller-ovs-59c8c\" (UID: \"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6\") " pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.742225 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-scripts\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.744284 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fd2a1d38-1169-4ab7-995f-eb6fb2f230a6-var-lib\") pod \"ovn-controller-ovs-59c8c\" (UID: \"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6\") " pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.744298 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fd2a1d38-1169-4ab7-995f-eb6fb2f230a6-var-log\") pod \"ovn-controller-ovs-59c8c\" (UID: \"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6\") " pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.744632 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-var-run-ovn\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.744676 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-var-run\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.744750 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fd2a1d38-1169-4ab7-995f-eb6fb2f230a6-var-run\") pod \"ovn-controller-ovs-59c8c\" (UID: \"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6\") " pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.744797 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-var-log-ovn\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.744809 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-scripts\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.744976 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fd2a1d38-1169-4ab7-995f-eb6fb2f230a6-etc-ovs\") pod \"ovn-controller-ovs-59c8c\" (UID: \"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6\") " pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.745929 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd2a1d38-1169-4ab7-995f-eb6fb2f230a6-scripts\") pod \"ovn-controller-ovs-59c8c\" (UID: \"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6\") " pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.753693 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-combined-ca-bundle\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.761297 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-ovn-controller-tls-certs\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.771873 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk8vf\" (UniqueName: \"kubernetes.io/projected/eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8-kube-api-access-fk8vf\") pod \"ovn-controller-hzbv4\" (UID: \"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8\") " pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.788297 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6s7j\" (UniqueName: \"kubernetes.io/projected/fd2a1d38-1169-4ab7-995f-eb6fb2f230a6-kube-api-access-n6s7j\") pod \"ovn-controller-ovs-59c8c\" (UID: \"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6\") " pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.859484 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:44 crc kubenswrapper[4808]: I1002 16:55:44.872065 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.148965 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.150915 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.154723 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.155011 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-skl6n" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.155383 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.155609 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.156041 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.177170 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.249632 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d642175e-72ec-466b-9273-b82beb5e99f5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.249704 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d642175e-72ec-466b-9273-b82beb5e99f5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.249729 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.249816 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d642175e-72ec-466b-9273-b82beb5e99f5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.249844 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d642175e-72ec-466b-9273-b82beb5e99f5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.249870 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d642175e-72ec-466b-9273-b82beb5e99f5-config\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.249893 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vljrw\" (UniqueName: \"kubernetes.io/projected/d642175e-72ec-466b-9273-b82beb5e99f5-kube-api-access-vljrw\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.249938 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d642175e-72ec-466b-9273-b82beb5e99f5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.351124 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d642175e-72ec-466b-9273-b82beb5e99f5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.351496 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d642175e-72ec-466b-9273-b82beb5e99f5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.351527 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.351615 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d642175e-72ec-466b-9273-b82beb5e99f5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.351641 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d642175e-72ec-466b-9273-b82beb5e99f5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.351668 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d642175e-72ec-466b-9273-b82beb5e99f5-config\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.351693 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vljrw\" (UniqueName: \"kubernetes.io/projected/d642175e-72ec-466b-9273-b82beb5e99f5-kube-api-access-vljrw\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.351767 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d642175e-72ec-466b-9273-b82beb5e99f5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.351954 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.352296 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d642175e-72ec-466b-9273-b82beb5e99f5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.353162 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d642175e-72ec-466b-9273-b82beb5e99f5-config\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.354095 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d642175e-72ec-466b-9273-b82beb5e99f5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.357981 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d642175e-72ec-466b-9273-b82beb5e99f5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.357996 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d642175e-72ec-466b-9273-b82beb5e99f5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.361989 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d642175e-72ec-466b-9273-b82beb5e99f5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.376813 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.377007 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vljrw\" (UniqueName: \"kubernetes.io/projected/d642175e-72ec-466b-9273-b82beb5e99f5-kube-api-access-vljrw\") pod \"ovsdbserver-nb-0\" (UID: \"d642175e-72ec-466b-9273-b82beb5e99f5\") " pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.474561 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 16:55:45 crc kubenswrapper[4808]: I1002 16:55:45.777848 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 16:55:46 crc kubenswrapper[4808]: W1002 16:55:46.290540 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb409b3d8_0ed6_49e9_a1ea_be6689e607ed.slice/crio-0cf96799e07035171b9c610fea6313b824d405d222f18bc02ec5266501087f96 WatchSource:0}: Error finding container 0cf96799e07035171b9c610fea6313b824d405d222f18bc02ec5266501087f96: Status 404 returned error can't find the container with id 0cf96799e07035171b9c610fea6313b824d405d222f18bc02ec5266501087f96 Oct 02 16:55:46 crc kubenswrapper[4808]: E1002 16:55:46.297067 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 02 16:55:46 crc kubenswrapper[4808]: E1002 16:55:46.297495 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6wwk4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-hkgw6_openstack(e85a2bd7-ef57-4edd-b643-0def775bd59b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 16:55:46 crc kubenswrapper[4808]: E1002 16:55:46.298724 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-hkgw6" podUID="e85a2bd7-ef57-4edd-b643-0def775bd59b" Oct 02 16:55:46 crc kubenswrapper[4808]: E1002 16:55:46.379955 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 02 16:55:46 crc kubenswrapper[4808]: E1002 16:55:46.380112 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4qjx8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-x6bs8_openstack(1636ee11-eba2-4ab3-9546-d824dbc5593b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 16:55:46 crc kubenswrapper[4808]: E1002 16:55:46.381631 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-x6bs8" podUID="1636ee11-eba2-4ab3-9546-d824dbc5593b" Oct 02 16:55:46 crc kubenswrapper[4808]: I1002 16:55:46.735799 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-xgpxx"] Oct 02 16:55:46 crc kubenswrapper[4808]: I1002 16:55:46.915384 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 16:55:46 crc kubenswrapper[4808]: I1002 16:55:46.918553 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:46 crc kubenswrapper[4808]: I1002 16:55:46.919958 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 16:55:46 crc kubenswrapper[4808]: I1002 16:55:46.921840 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 02 16:55:46 crc kubenswrapper[4808]: I1002 16:55:46.921851 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-q8q7n" Oct 02 16:55:46 crc kubenswrapper[4808]: I1002 16:55:46.921896 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 02 16:55:46 crc kubenswrapper[4808]: I1002 16:55:46.922190 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 02 16:55:46 crc kubenswrapper[4808]: I1002 16:55:46.979536 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:46 crc kubenswrapper[4808]: I1002 16:55:46.979586 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:46 crc kubenswrapper[4808]: I1002 16:55:46.979661 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-config\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:46 crc kubenswrapper[4808]: I1002 16:55:46.979677 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnscs\" (UniqueName: \"kubernetes.io/projected/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-kube-api-access-gnscs\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:46 crc kubenswrapper[4808]: I1002 16:55:46.979691 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:46 crc kubenswrapper[4808]: I1002 16:55:46.979891 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:46 crc kubenswrapper[4808]: I1002 16:55:46.979931 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:46 crc kubenswrapper[4808]: I1002 16:55:46.979958 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.081606 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-config\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.081648 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnscs\" (UniqueName: \"kubernetes.io/projected/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-kube-api-access-gnscs\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.081670 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.081753 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.081778 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.081804 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.081840 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.081865 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.082318 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.082997 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-config\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.083735 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.084414 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.088263 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.088467 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.088597 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.098260 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnscs\" (UniqueName: \"kubernetes.io/projected/dbb01d1d-5fc2-4d1c-8fa3-5575de879622-kube-api-access-gnscs\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.115672 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dbb01d1d-5fc2-4d1c-8fa3-5575de879622\") " pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.131215 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-hzbv4"] Oct 02 16:55:47 crc kubenswrapper[4808]: W1002 16:55:47.132363 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeafc4ad2_d1ad_446a_ae4e_4fd250aaa5f8.slice/crio-a9510af9adc829832fbaba491e92d6ba02de40f1191eb47bb474426918d000fc WatchSource:0}: Error finding container a9510af9adc829832fbaba491e92d6ba02de40f1191eb47bb474426918d000fc: Status 404 returned error can't find the container with id a9510af9adc829832fbaba491e92d6ba02de40f1191eb47bb474426918d000fc Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.190359 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.203359 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 16:55:47 crc kubenswrapper[4808]: W1002 16:55:47.203381 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf71818d6_1ebf_4838_b871_8873c107b325.slice/crio-3fa89b1cc3dfc663ebc7dec8bcfefa85feb57fd1577efc4daeff8f04db62989f WatchSource:0}: Error finding container 3fa89b1cc3dfc663ebc7dec8bcfefa85feb57fd1577efc4daeff8f04db62989f: Status 404 returned error can't find the container with id 3fa89b1cc3dfc663ebc7dec8bcfefa85feb57fd1577efc4daeff8f04db62989f Oct 02 16:55:47 crc kubenswrapper[4808]: W1002 16:55:47.204469 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8509820_f0d7_417b_92d4_1be8b7b714ef.slice/crio-708b68b1273cd9b54ff100158af93880f3c9600cc95c5157aa46d1a1872da4c7 WatchSource:0}: Error finding container 708b68b1273cd9b54ff100158af93880f3c9600cc95c5157aa46d1a1872da4c7: Status 404 returned error can't find the container with id 708b68b1273cd9b54ff100158af93880f3c9600cc95c5157aa46d1a1872da4c7 Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.213523 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.225969 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.263702 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.265807 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 16:55:47 crc kubenswrapper[4808]: W1002 16:55:47.266546 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf18f563e_79e6_4c33_87a5_91f7c27bf181.slice/crio-d02e37733c60add96e9eafa89bb29e873dfd6102c813fe0f86fbaed5ec4cfb96 WatchSource:0}: Error finding container d02e37733c60add96e9eafa89bb29e873dfd6102c813fe0f86fbaed5ec4cfb96: Status 404 returned error can't find the container with id d02e37733c60add96e9eafa89bb29e873dfd6102c813fe0f86fbaed5ec4cfb96 Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.282009 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f71818d6-1ebf-4838-b871-8873c107b325","Type":"ContainerStarted","Data":"3fa89b1cc3dfc663ebc7dec8bcfefa85feb57fd1577efc4daeff8f04db62989f"} Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.282725 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f18f563e-79e6-4c33-87a5-91f7c27bf181","Type":"ContainerStarted","Data":"d02e37733c60add96e9eafa89bb29e873dfd6102c813fe0f86fbaed5ec4cfb96"} Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.285782 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b409b3d8-0ed6-49e9-a1ea-be6689e607ed","Type":"ContainerStarted","Data":"0cf96799e07035171b9c610fea6313b824d405d222f18bc02ec5266501087f96"} Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.288208 4808 generic.go:334] "Generic (PLEG): container finished" podID="b11c2afc-30b4-4795-870c-1b710a87c054" containerID="016aac327c423dc735cc3a5f96c1e06e750824a7388fedaf78ba85b2cbe30b93" exitCode=0 Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.288366 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" event={"ID":"b11c2afc-30b4-4795-870c-1b710a87c054","Type":"ContainerDied","Data":"016aac327c423dc735cc3a5f96c1e06e750824a7388fedaf78ba85b2cbe30b93"} Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.290844 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a8509820-f0d7-417b-92d4-1be8b7b714ef","Type":"ContainerStarted","Data":"708b68b1273cd9b54ff100158af93880f3c9600cc95c5157aa46d1a1872da4c7"} Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.292501 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hzbv4" event={"ID":"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8","Type":"ContainerStarted","Data":"a9510af9adc829832fbaba491e92d6ba02de40f1191eb47bb474426918d000fc"} Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.293846 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"8b0cce8a-5f59-490d-87e6-6186890cd2a8","Type":"ContainerStarted","Data":"b361e8c18df147148dbdaa7005f9210a56a823463f0a9a07f790bb909bf29f66"} Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.295606 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"500b4d21-bf0e-4763-91bd-7c32bf13a064","Type":"ContainerStarted","Data":"843535a82aa883ed71bf0bf2cac1ae28556d4b91330da760644daec3f7b372a1"} Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.298839 4808 generic.go:334] "Generic (PLEG): container finished" podID="45c90d4a-2e61-431a-97b4-900c0ceb6493" containerID="6f3066add9aae146979e59164ea6c6224346e492c34c1772e8493e7651d78036" exitCode=0 Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.298978 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" event={"ID":"45c90d4a-2e61-431a-97b4-900c0ceb6493","Type":"ContainerDied","Data":"6f3066add9aae146979e59164ea6c6224346e492c34c1772e8493e7651d78036"} Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.299000 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" event={"ID":"45c90d4a-2e61-431a-97b4-900c0ceb6493","Type":"ContainerStarted","Data":"fac50baba4bb7f5df8fb1f22ab82167eb2d30460de34191414022f896e0c8b6d"} Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.325116 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 16:55:47 crc kubenswrapper[4808]: W1002 16:55:47.343649 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd642175e_72ec_466b_9273_b82beb5e99f5.slice/crio-45458db08bf4b3776cae87b6bc247780f1b10e714b692171334991e88d181459 WatchSource:0}: Error finding container 45458db08bf4b3776cae87b6bc247780f1b10e714b692171334991e88d181459: Status 404 returned error can't find the container with id 45458db08bf4b3776cae87b6bc247780f1b10e714b692171334991e88d181459 Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.448834 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-59c8c"] Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.669735 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-x6bs8" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.716375 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-hkgw6" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.793306 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wwk4\" (UniqueName: \"kubernetes.io/projected/e85a2bd7-ef57-4edd-b643-0def775bd59b-kube-api-access-6wwk4\") pod \"e85a2bd7-ef57-4edd-b643-0def775bd59b\" (UID: \"e85a2bd7-ef57-4edd-b643-0def775bd59b\") " Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.793369 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e85a2bd7-ef57-4edd-b643-0def775bd59b-dns-svc\") pod \"e85a2bd7-ef57-4edd-b643-0def775bd59b\" (UID: \"e85a2bd7-ef57-4edd-b643-0def775bd59b\") " Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.793422 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1636ee11-eba2-4ab3-9546-d824dbc5593b-config\") pod \"1636ee11-eba2-4ab3-9546-d824dbc5593b\" (UID: \"1636ee11-eba2-4ab3-9546-d824dbc5593b\") " Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.793446 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qjx8\" (UniqueName: \"kubernetes.io/projected/1636ee11-eba2-4ab3-9546-d824dbc5593b-kube-api-access-4qjx8\") pod \"1636ee11-eba2-4ab3-9546-d824dbc5593b\" (UID: \"1636ee11-eba2-4ab3-9546-d824dbc5593b\") " Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.793467 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e85a2bd7-ef57-4edd-b643-0def775bd59b-config\") pod \"e85a2bd7-ef57-4edd-b643-0def775bd59b\" (UID: \"e85a2bd7-ef57-4edd-b643-0def775bd59b\") " Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.794090 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e85a2bd7-ef57-4edd-b643-0def775bd59b-config" (OuterVolumeSpecName: "config") pod "e85a2bd7-ef57-4edd-b643-0def775bd59b" (UID: "e85a2bd7-ef57-4edd-b643-0def775bd59b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.794130 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e85a2bd7-ef57-4edd-b643-0def775bd59b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e85a2bd7-ef57-4edd-b643-0def775bd59b" (UID: "e85a2bd7-ef57-4edd-b643-0def775bd59b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.794152 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1636ee11-eba2-4ab3-9546-d824dbc5593b-config" (OuterVolumeSpecName: "config") pod "1636ee11-eba2-4ab3-9546-d824dbc5593b" (UID: "1636ee11-eba2-4ab3-9546-d824dbc5593b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.799000 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1636ee11-eba2-4ab3-9546-d824dbc5593b-kube-api-access-4qjx8" (OuterVolumeSpecName: "kube-api-access-4qjx8") pod "1636ee11-eba2-4ab3-9546-d824dbc5593b" (UID: "1636ee11-eba2-4ab3-9546-d824dbc5593b"). InnerVolumeSpecName "kube-api-access-4qjx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.800365 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e85a2bd7-ef57-4edd-b643-0def775bd59b-kube-api-access-6wwk4" (OuterVolumeSpecName: "kube-api-access-6wwk4") pod "e85a2bd7-ef57-4edd-b643-0def775bd59b" (UID: "e85a2bd7-ef57-4edd-b643-0def775bd59b"). InnerVolumeSpecName "kube-api-access-6wwk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.899003 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e85a2bd7-ef57-4edd-b643-0def775bd59b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.899033 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1636ee11-eba2-4ab3-9546-d824dbc5593b-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.899044 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qjx8\" (UniqueName: \"kubernetes.io/projected/1636ee11-eba2-4ab3-9546-d824dbc5593b-kube-api-access-4qjx8\") on node \"crc\" DevicePath \"\"" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.899053 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e85a2bd7-ef57-4edd-b643-0def775bd59b-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.899062 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wwk4\" (UniqueName: \"kubernetes.io/projected/e85a2bd7-ef57-4edd-b643-0def775bd59b-kube-api-access-6wwk4\") on node \"crc\" DevicePath \"\"" Oct 02 16:55:47 crc kubenswrapper[4808]: I1002 16:55:47.928179 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 16:55:48 crc kubenswrapper[4808]: I1002 16:55:48.307096 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d642175e-72ec-466b-9273-b82beb5e99f5","Type":"ContainerStarted","Data":"45458db08bf4b3776cae87b6bc247780f1b10e714b692171334991e88d181459"} Oct 02 16:55:48 crc kubenswrapper[4808]: I1002 16:55:48.309078 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-x6bs8" event={"ID":"1636ee11-eba2-4ab3-9546-d824dbc5593b","Type":"ContainerDied","Data":"57c16fd7f7cded477ce0ce4523173eeea7ebd27bb4b88edac5327e0009863fc4"} Oct 02 16:55:48 crc kubenswrapper[4808]: I1002 16:55:48.309143 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-x6bs8" Oct 02 16:55:48 crc kubenswrapper[4808]: I1002 16:55:48.345863 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" event={"ID":"45c90d4a-2e61-431a-97b4-900c0ceb6493","Type":"ContainerStarted","Data":"7a105c68d069497311e69ee8c3f8111d30a9957c091c2b271a728e67b7239eba"} Oct 02 16:55:48 crc kubenswrapper[4808]: I1002 16:55:48.346602 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" Oct 02 16:55:48 crc kubenswrapper[4808]: I1002 16:55:48.349041 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-59c8c" event={"ID":"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6","Type":"ContainerStarted","Data":"94771520b7ba11c3414eab90fa26adf85dcc52e6c876f17a2caa58fd570acc01"} Oct 02 16:55:48 crc kubenswrapper[4808]: I1002 16:55:48.365443 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" event={"ID":"b11c2afc-30b4-4795-870c-1b710a87c054","Type":"ContainerStarted","Data":"db1d6e4924f83326bd8279f44b2ef8e5e0ec4d886f47e609dee1c90ed4289d09"} Oct 02 16:55:48 crc kubenswrapper[4808]: I1002 16:55:48.366037 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" Oct 02 16:55:48 crc kubenswrapper[4808]: I1002 16:55:48.376929 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"dbb01d1d-5fc2-4d1c-8fa3-5575de879622","Type":"ContainerStarted","Data":"1831a4f01645e97c52ffa59d08e463f84c7b0329904c7c81874f9a99d770676b"} Oct 02 16:55:48 crc kubenswrapper[4808]: I1002 16:55:48.378748 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-x6bs8"] Oct 02 16:55:48 crc kubenswrapper[4808]: I1002 16:55:48.386026 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-hkgw6" event={"ID":"e85a2bd7-ef57-4edd-b643-0def775bd59b","Type":"ContainerDied","Data":"bb6742ce0ca1834d0e8ee6737fbf8fe679728dbd6d16547be4689afc07c4595d"} Oct 02 16:55:48 crc kubenswrapper[4808]: I1002 16:55:48.386126 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-hkgw6" Oct 02 16:55:48 crc kubenswrapper[4808]: I1002 16:55:48.386753 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-x6bs8"] Oct 02 16:55:48 crc kubenswrapper[4808]: I1002 16:55:48.422039 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" podStartSLOduration=15.422022886 podStartE2EDuration="15.422022886s" podCreationTimestamp="2025-10-02 16:55:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:55:48.420111075 +0000 UTC m=+915.745640075" watchObservedRunningTime="2025-10-02 16:55:48.422022886 +0000 UTC m=+915.747551896" Oct 02 16:55:48 crc kubenswrapper[4808]: I1002 16:55:48.422554 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" podStartSLOduration=6.51465261 podStartE2EDuration="15.422548031s" podCreationTimestamp="2025-10-02 16:55:33 +0000 UTC" firstStartedPulling="2025-10-02 16:55:37.548147335 +0000 UTC m=+904.873676335" lastFinishedPulling="2025-10-02 16:55:46.456042756 +0000 UTC m=+913.781571756" observedRunningTime="2025-10-02 16:55:48.405883539 +0000 UTC m=+915.731412539" watchObservedRunningTime="2025-10-02 16:55:48.422548031 +0000 UTC m=+915.748077031" Oct 02 16:55:48 crc kubenswrapper[4808]: I1002 16:55:48.452644 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-hkgw6"] Oct 02 16:55:48 crc kubenswrapper[4808]: I1002 16:55:48.457777 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-hkgw6"] Oct 02 16:55:49 crc kubenswrapper[4808]: I1002 16:55:49.408919 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1636ee11-eba2-4ab3-9546-d824dbc5593b" path="/var/lib/kubelet/pods/1636ee11-eba2-4ab3-9546-d824dbc5593b/volumes" Oct 02 16:55:49 crc kubenswrapper[4808]: I1002 16:55:49.409311 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e85a2bd7-ef57-4edd-b643-0def775bd59b" path="/var/lib/kubelet/pods/e85a2bd7-ef57-4edd-b643-0def775bd59b/volumes" Oct 02 16:55:54 crc kubenswrapper[4808]: I1002 16:55:54.001630 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" Oct 02 16:55:54 crc kubenswrapper[4808]: I1002 16:55:54.319994 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" Oct 02 16:55:54 crc kubenswrapper[4808]: I1002 16:55:54.381224 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-t2nm2"] Oct 02 16:55:54 crc kubenswrapper[4808]: I1002 16:55:54.447139 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" podUID="b11c2afc-30b4-4795-870c-1b710a87c054" containerName="dnsmasq-dns" containerID="cri-o://db1d6e4924f83326bd8279f44b2ef8e5e0ec4d886f47e609dee1c90ed4289d09" gracePeriod=10 Oct 02 16:55:55 crc kubenswrapper[4808]: I1002 16:55:55.427344 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" Oct 02 16:55:55 crc kubenswrapper[4808]: I1002 16:55:55.477688 4808 generic.go:334] "Generic (PLEG): container finished" podID="b11c2afc-30b4-4795-870c-1b710a87c054" containerID="db1d6e4924f83326bd8279f44b2ef8e5e0ec4d886f47e609dee1c90ed4289d09" exitCode=0 Oct 02 16:55:55 crc kubenswrapper[4808]: I1002 16:55:55.477726 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" event={"ID":"b11c2afc-30b4-4795-870c-1b710a87c054","Type":"ContainerDied","Data":"db1d6e4924f83326bd8279f44b2ef8e5e0ec4d886f47e609dee1c90ed4289d09"} Oct 02 16:55:55 crc kubenswrapper[4808]: I1002 16:55:55.477752 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" event={"ID":"b11c2afc-30b4-4795-870c-1b710a87c054","Type":"ContainerDied","Data":"433ff68fa06af4af4b1df7ad203c2ba462f930731ae2a4faa5de8a0ef88e1a5a"} Oct 02 16:55:55 crc kubenswrapper[4808]: I1002 16:55:55.477767 4808 scope.go:117] "RemoveContainer" containerID="db1d6e4924f83326bd8279f44b2ef8e5e0ec4d886f47e609dee1c90ed4289d09" Oct 02 16:55:55 crc kubenswrapper[4808]: I1002 16:55:55.477871 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-t2nm2" Oct 02 16:55:55 crc kubenswrapper[4808]: I1002 16:55:55.549995 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b11c2afc-30b4-4795-870c-1b710a87c054-dns-svc\") pod \"b11c2afc-30b4-4795-870c-1b710a87c054\" (UID: \"b11c2afc-30b4-4795-870c-1b710a87c054\") " Oct 02 16:55:55 crc kubenswrapper[4808]: I1002 16:55:55.550035 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmn2m\" (UniqueName: \"kubernetes.io/projected/b11c2afc-30b4-4795-870c-1b710a87c054-kube-api-access-bmn2m\") pod \"b11c2afc-30b4-4795-870c-1b710a87c054\" (UID: \"b11c2afc-30b4-4795-870c-1b710a87c054\") " Oct 02 16:55:55 crc kubenswrapper[4808]: I1002 16:55:55.550133 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b11c2afc-30b4-4795-870c-1b710a87c054-config\") pod \"b11c2afc-30b4-4795-870c-1b710a87c054\" (UID: \"b11c2afc-30b4-4795-870c-1b710a87c054\") " Oct 02 16:55:55 crc kubenswrapper[4808]: I1002 16:55:55.553181 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11c2afc-30b4-4795-870c-1b710a87c054-kube-api-access-bmn2m" (OuterVolumeSpecName: "kube-api-access-bmn2m") pod "b11c2afc-30b4-4795-870c-1b710a87c054" (UID: "b11c2afc-30b4-4795-870c-1b710a87c054"). InnerVolumeSpecName "kube-api-access-bmn2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:55:55 crc kubenswrapper[4808]: I1002 16:55:55.604500 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b11c2afc-30b4-4795-870c-1b710a87c054-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b11c2afc-30b4-4795-870c-1b710a87c054" (UID: "b11c2afc-30b4-4795-870c-1b710a87c054"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:55:55 crc kubenswrapper[4808]: I1002 16:55:55.612379 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b11c2afc-30b4-4795-870c-1b710a87c054-config" (OuterVolumeSpecName: "config") pod "b11c2afc-30b4-4795-870c-1b710a87c054" (UID: "b11c2afc-30b4-4795-870c-1b710a87c054"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:55:55 crc kubenswrapper[4808]: I1002 16:55:55.652887 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b11c2afc-30b4-4795-870c-1b710a87c054-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 16:55:55 crc kubenswrapper[4808]: I1002 16:55:55.652927 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmn2m\" (UniqueName: \"kubernetes.io/projected/b11c2afc-30b4-4795-870c-1b710a87c054-kube-api-access-bmn2m\") on node \"crc\" DevicePath \"\"" Oct 02 16:55:55 crc kubenswrapper[4808]: I1002 16:55:55.652944 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b11c2afc-30b4-4795-870c-1b710a87c054-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:55:55 crc kubenswrapper[4808]: I1002 16:55:55.810627 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-t2nm2"] Oct 02 16:55:55 crc kubenswrapper[4808]: I1002 16:55:55.815717 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-t2nm2"] Oct 02 16:55:55 crc kubenswrapper[4808]: I1002 16:55:55.954885 4808 scope.go:117] "RemoveContainer" containerID="016aac327c423dc735cc3a5f96c1e06e750824a7388fedaf78ba85b2cbe30b93" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.127275 4808 scope.go:117] "RemoveContainer" containerID="db1d6e4924f83326bd8279f44b2ef8e5e0ec4d886f47e609dee1c90ed4289d09" Oct 02 16:55:56 crc kubenswrapper[4808]: E1002 16:55:56.128135 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db1d6e4924f83326bd8279f44b2ef8e5e0ec4d886f47e609dee1c90ed4289d09\": container with ID starting with db1d6e4924f83326bd8279f44b2ef8e5e0ec4d886f47e609dee1c90ed4289d09 not found: ID does not exist" containerID="db1d6e4924f83326bd8279f44b2ef8e5e0ec4d886f47e609dee1c90ed4289d09" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.128209 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db1d6e4924f83326bd8279f44b2ef8e5e0ec4d886f47e609dee1c90ed4289d09"} err="failed to get container status \"db1d6e4924f83326bd8279f44b2ef8e5e0ec4d886f47e609dee1c90ed4289d09\": rpc error: code = NotFound desc = could not find container \"db1d6e4924f83326bd8279f44b2ef8e5e0ec4d886f47e609dee1c90ed4289d09\": container with ID starting with db1d6e4924f83326bd8279f44b2ef8e5e0ec4d886f47e609dee1c90ed4289d09 not found: ID does not exist" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.128269 4808 scope.go:117] "RemoveContainer" containerID="016aac327c423dc735cc3a5f96c1e06e750824a7388fedaf78ba85b2cbe30b93" Oct 02 16:55:56 crc kubenswrapper[4808]: E1002 16:55:56.128908 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"016aac327c423dc735cc3a5f96c1e06e750824a7388fedaf78ba85b2cbe30b93\": container with ID starting with 016aac327c423dc735cc3a5f96c1e06e750824a7388fedaf78ba85b2cbe30b93 not found: ID does not exist" containerID="016aac327c423dc735cc3a5f96c1e06e750824a7388fedaf78ba85b2cbe30b93" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.128942 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"016aac327c423dc735cc3a5f96c1e06e750824a7388fedaf78ba85b2cbe30b93"} err="failed to get container status \"016aac327c423dc735cc3a5f96c1e06e750824a7388fedaf78ba85b2cbe30b93\": rpc error: code = NotFound desc = could not find container \"016aac327c423dc735cc3a5f96c1e06e750824a7388fedaf78ba85b2cbe30b93\": container with ID starting with 016aac327c423dc735cc3a5f96c1e06e750824a7388fedaf78ba85b2cbe30b93 not found: ID does not exist" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.332623 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-8d9tc"] Oct 02 16:55:56 crc kubenswrapper[4808]: E1002 16:55:56.332981 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b11c2afc-30b4-4795-870c-1b710a87c054" containerName="dnsmasq-dns" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.333001 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b11c2afc-30b4-4795-870c-1b710a87c054" containerName="dnsmasq-dns" Oct 02 16:55:56 crc kubenswrapper[4808]: E1002 16:55:56.333048 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b11c2afc-30b4-4795-870c-1b710a87c054" containerName="init" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.333058 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b11c2afc-30b4-4795-870c-1b710a87c054" containerName="init" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.333325 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="b11c2afc-30b4-4795-870c-1b710a87c054" containerName="dnsmasq-dns" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.333958 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-8d9tc" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.341265 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.357924 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-8d9tc"] Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.461551 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-qgt5h"] Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.466243 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-qgt5h" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.467245 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebecde28-66be-4caa-87d4-779d8605da87-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-8d9tc\" (UID: \"ebecde28-66be-4caa-87d4-779d8605da87\") " pod="openstack/ovn-controller-metrics-8d9tc" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.467302 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebecde28-66be-4caa-87d4-779d8605da87-config\") pod \"ovn-controller-metrics-8d9tc\" (UID: \"ebecde28-66be-4caa-87d4-779d8605da87\") " pod="openstack/ovn-controller-metrics-8d9tc" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.467323 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebecde28-66be-4caa-87d4-779d8605da87-combined-ca-bundle\") pod \"ovn-controller-metrics-8d9tc\" (UID: \"ebecde28-66be-4caa-87d4-779d8605da87\") " pod="openstack/ovn-controller-metrics-8d9tc" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.467472 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ebecde28-66be-4caa-87d4-779d8605da87-ovn-rundir\") pod \"ovn-controller-metrics-8d9tc\" (UID: \"ebecde28-66be-4caa-87d4-779d8605da87\") " pod="openstack/ovn-controller-metrics-8d9tc" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.467638 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-995jz\" (UniqueName: \"kubernetes.io/projected/ebecde28-66be-4caa-87d4-779d8605da87-kube-api-access-995jz\") pod \"ovn-controller-metrics-8d9tc\" (UID: \"ebecde28-66be-4caa-87d4-779d8605da87\") " pod="openstack/ovn-controller-metrics-8d9tc" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.467678 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ebecde28-66be-4caa-87d4-779d8605da87-ovs-rundir\") pod \"ovn-controller-metrics-8d9tc\" (UID: \"ebecde28-66be-4caa-87d4-779d8605da87\") " pod="openstack/ovn-controller-metrics-8d9tc" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.471634 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.474043 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-qgt5h"] Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.507360 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a8509820-f0d7-417b-92d4-1be8b7b714ef","Type":"ContainerStarted","Data":"fe6b573c8f477b53c3c392d96d8d584666a437a31a77b0a91bc86d96a6fc55c0"} Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.516461 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.527416 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"dbb01d1d-5fc2-4d1c-8fa3-5575de879622","Type":"ContainerStarted","Data":"ae2b91f64dc6c858d6b70f7dabf759c62fcf36a41c30e08cbf1f9ee4efc66d1b"} Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.534206 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"8b0cce8a-5f59-490d-87e6-6186890cd2a8","Type":"ContainerStarted","Data":"38aa967d58eaf87a9ce8c37e611409e8fbe4794a53123b289fa1b9f0cd27038d"} Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.534895 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.536696 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d642175e-72ec-466b-9273-b82beb5e99f5","Type":"ContainerStarted","Data":"e3fd8d48873ebe53a6a127bb2a9468547e27d67f543d7345f4e0328ab024d212"} Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.559600 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=7.760501658 podStartE2EDuration="16.559581455s" podCreationTimestamp="2025-10-02 16:55:40 +0000 UTC" firstStartedPulling="2025-10-02 16:55:47.27584523 +0000 UTC m=+914.601374250" lastFinishedPulling="2025-10-02 16:55:56.074925027 +0000 UTC m=+923.400454047" observedRunningTime="2025-10-02 16:55:56.554457866 +0000 UTC m=+923.879986866" watchObservedRunningTime="2025-10-02 16:55:56.559581455 +0000 UTC m=+923.885110455" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.569803 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d38d552b-cc48-4ad2-b6df-873129d015e8-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-qgt5h\" (UID: \"d38d552b-cc48-4ad2-b6df-873129d015e8\") " pod="openstack/dnsmasq-dns-7fd796d7df-qgt5h" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.569885 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebecde28-66be-4caa-87d4-779d8605da87-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-8d9tc\" (UID: \"ebecde28-66be-4caa-87d4-779d8605da87\") " pod="openstack/ovn-controller-metrics-8d9tc" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.569927 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d38d552b-cc48-4ad2-b6df-873129d015e8-config\") pod \"dnsmasq-dns-7fd796d7df-qgt5h\" (UID: \"d38d552b-cc48-4ad2-b6df-873129d015e8\") " pod="openstack/dnsmasq-dns-7fd796d7df-qgt5h" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.569944 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebecde28-66be-4caa-87d4-779d8605da87-config\") pod \"ovn-controller-metrics-8d9tc\" (UID: \"ebecde28-66be-4caa-87d4-779d8605da87\") " pod="openstack/ovn-controller-metrics-8d9tc" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.569960 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebecde28-66be-4caa-87d4-779d8605da87-combined-ca-bundle\") pod \"ovn-controller-metrics-8d9tc\" (UID: \"ebecde28-66be-4caa-87d4-779d8605da87\") " pod="openstack/ovn-controller-metrics-8d9tc" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.570011 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ebecde28-66be-4caa-87d4-779d8605da87-ovn-rundir\") pod \"ovn-controller-metrics-8d9tc\" (UID: \"ebecde28-66be-4caa-87d4-779d8605da87\") " pod="openstack/ovn-controller-metrics-8d9tc" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.570027 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d38d552b-cc48-4ad2-b6df-873129d015e8-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-qgt5h\" (UID: \"d38d552b-cc48-4ad2-b6df-873129d015e8\") " pod="openstack/dnsmasq-dns-7fd796d7df-qgt5h" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.570042 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m24lp\" (UniqueName: \"kubernetes.io/projected/d38d552b-cc48-4ad2-b6df-873129d015e8-kube-api-access-m24lp\") pod \"dnsmasq-dns-7fd796d7df-qgt5h\" (UID: \"d38d552b-cc48-4ad2-b6df-873129d015e8\") " pod="openstack/dnsmasq-dns-7fd796d7df-qgt5h" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.570086 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-995jz\" (UniqueName: \"kubernetes.io/projected/ebecde28-66be-4caa-87d4-779d8605da87-kube-api-access-995jz\") pod \"ovn-controller-metrics-8d9tc\" (UID: \"ebecde28-66be-4caa-87d4-779d8605da87\") " pod="openstack/ovn-controller-metrics-8d9tc" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.570120 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ebecde28-66be-4caa-87d4-779d8605da87-ovs-rundir\") pod \"ovn-controller-metrics-8d9tc\" (UID: \"ebecde28-66be-4caa-87d4-779d8605da87\") " pod="openstack/ovn-controller-metrics-8d9tc" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.570455 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ebecde28-66be-4caa-87d4-779d8605da87-ovs-rundir\") pod \"ovn-controller-metrics-8d9tc\" (UID: \"ebecde28-66be-4caa-87d4-779d8605da87\") " pod="openstack/ovn-controller-metrics-8d9tc" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.573806 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ebecde28-66be-4caa-87d4-779d8605da87-ovn-rundir\") pod \"ovn-controller-metrics-8d9tc\" (UID: \"ebecde28-66be-4caa-87d4-779d8605da87\") " pod="openstack/ovn-controller-metrics-8d9tc" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.576451 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebecde28-66be-4caa-87d4-779d8605da87-config\") pod \"ovn-controller-metrics-8d9tc\" (UID: \"ebecde28-66be-4caa-87d4-779d8605da87\") " pod="openstack/ovn-controller-metrics-8d9tc" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.578787 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebecde28-66be-4caa-87d4-779d8605da87-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-8d9tc\" (UID: \"ebecde28-66be-4caa-87d4-779d8605da87\") " pod="openstack/ovn-controller-metrics-8d9tc" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.581094 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebecde28-66be-4caa-87d4-779d8605da87-combined-ca-bundle\") pod \"ovn-controller-metrics-8d9tc\" (UID: \"ebecde28-66be-4caa-87d4-779d8605da87\") " pod="openstack/ovn-controller-metrics-8d9tc" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.595296 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-qgt5h"] Oct 02 16:55:56 crc kubenswrapper[4808]: E1002 16:55:56.596035 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-m24lp ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7fd796d7df-qgt5h" podUID="d38d552b-cc48-4ad2-b6df-873129d015e8" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.596447 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-995jz\" (UniqueName: \"kubernetes.io/projected/ebecde28-66be-4caa-87d4-779d8605da87-kube-api-access-995jz\") pod \"ovn-controller-metrics-8d9tc\" (UID: \"ebecde28-66be-4caa-87d4-779d8605da87\") " pod="openstack/ovn-controller-metrics-8d9tc" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.599561 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=10.634250415 podStartE2EDuration="18.599542567s" podCreationTimestamp="2025-10-02 16:55:38 +0000 UTC" firstStartedPulling="2025-10-02 16:55:47.209058551 +0000 UTC m=+914.534587561" lastFinishedPulling="2025-10-02 16:55:55.174350673 +0000 UTC m=+922.499879713" observedRunningTime="2025-10-02 16:55:56.573177783 +0000 UTC m=+923.898706783" watchObservedRunningTime="2025-10-02 16:55:56.599542567 +0000 UTC m=+923.925071557" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.653109 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-l556v"] Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.654735 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.657979 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.658442 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-8d9tc" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.658834 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-l556v"] Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.675459 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d38d552b-cc48-4ad2-b6df-873129d015e8-config\") pod \"dnsmasq-dns-7fd796d7df-qgt5h\" (UID: \"d38d552b-cc48-4ad2-b6df-873129d015e8\") " pod="openstack/dnsmasq-dns-7fd796d7df-qgt5h" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.675722 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d38d552b-cc48-4ad2-b6df-873129d015e8-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-qgt5h\" (UID: \"d38d552b-cc48-4ad2-b6df-873129d015e8\") " pod="openstack/dnsmasq-dns-7fd796d7df-qgt5h" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.675801 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m24lp\" (UniqueName: \"kubernetes.io/projected/d38d552b-cc48-4ad2-b6df-873129d015e8-kube-api-access-m24lp\") pod \"dnsmasq-dns-7fd796d7df-qgt5h\" (UID: \"d38d552b-cc48-4ad2-b6df-873129d015e8\") " pod="openstack/dnsmasq-dns-7fd796d7df-qgt5h" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.675936 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d38d552b-cc48-4ad2-b6df-873129d015e8-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-qgt5h\" (UID: \"d38d552b-cc48-4ad2-b6df-873129d015e8\") " pod="openstack/dnsmasq-dns-7fd796d7df-qgt5h" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.676491 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d38d552b-cc48-4ad2-b6df-873129d015e8-config\") pod \"dnsmasq-dns-7fd796d7df-qgt5h\" (UID: \"d38d552b-cc48-4ad2-b6df-873129d015e8\") " pod="openstack/dnsmasq-dns-7fd796d7df-qgt5h" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.677680 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d38d552b-cc48-4ad2-b6df-873129d015e8-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-qgt5h\" (UID: \"d38d552b-cc48-4ad2-b6df-873129d015e8\") " pod="openstack/dnsmasq-dns-7fd796d7df-qgt5h" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.677790 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d38d552b-cc48-4ad2-b6df-873129d015e8-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-qgt5h\" (UID: \"d38d552b-cc48-4ad2-b6df-873129d015e8\") " pod="openstack/dnsmasq-dns-7fd796d7df-qgt5h" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.697098 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m24lp\" (UniqueName: \"kubernetes.io/projected/d38d552b-cc48-4ad2-b6df-873129d015e8-kube-api-access-m24lp\") pod \"dnsmasq-dns-7fd796d7df-qgt5h\" (UID: \"d38d552b-cc48-4ad2-b6df-873129d015e8\") " pod="openstack/dnsmasq-dns-7fd796d7df-qgt5h" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.777284 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-config\") pod \"dnsmasq-dns-86db49b7ff-l556v\" (UID: \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\") " pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.777687 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-l556v\" (UID: \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\") " pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.777709 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-l556v\" (UID: \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\") " pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.777731 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbvcz\" (UniqueName: \"kubernetes.io/projected/181f7b50-6e00-4e8c-bdde-9bbff23c3306-kube-api-access-sbvcz\") pod \"dnsmasq-dns-86db49b7ff-l556v\" (UID: \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\") " pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.777748 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-l556v\" (UID: \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\") " pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.879378 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-config\") pod \"dnsmasq-dns-86db49b7ff-l556v\" (UID: \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\") " pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.879460 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-l556v\" (UID: \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\") " pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.879482 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-l556v\" (UID: \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\") " pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.879505 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbvcz\" (UniqueName: \"kubernetes.io/projected/181f7b50-6e00-4e8c-bdde-9bbff23c3306-kube-api-access-sbvcz\") pod \"dnsmasq-dns-86db49b7ff-l556v\" (UID: \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\") " pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.879523 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-l556v\" (UID: \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\") " pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.880809 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-l556v\" (UID: \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\") " pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.880874 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-config\") pod \"dnsmasq-dns-86db49b7ff-l556v\" (UID: \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\") " pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.880959 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-l556v\" (UID: \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\") " pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.881005 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-l556v\" (UID: \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\") " pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.899477 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbvcz\" (UniqueName: \"kubernetes.io/projected/181f7b50-6e00-4e8c-bdde-9bbff23c3306-kube-api-access-sbvcz\") pod \"dnsmasq-dns-86db49b7ff-l556v\" (UID: \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\") " pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:55:56 crc kubenswrapper[4808]: I1002 16:55:56.984945 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.097620 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-8d9tc"] Oct 02 16:55:57 crc kubenswrapper[4808]: W1002 16:55:57.110101 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebecde28_66be_4caa_87d4_779d8605da87.slice/crio-a8cc68871a8693e11330ac5365a276ea38af84f53607ee5d9bf34f7d26e7fbd9 WatchSource:0}: Error finding container a8cc68871a8693e11330ac5365a276ea38af84f53607ee5d9bf34f7d26e7fbd9: Status 404 returned error can't find the container with id a8cc68871a8693e11330ac5365a276ea38af84f53607ee5d9bf34f7d26e7fbd9 Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.413506 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11c2afc-30b4-4795-870c-1b710a87c054" path="/var/lib/kubelet/pods/b11c2afc-30b4-4795-870c-1b710a87c054/volumes" Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.423925 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-l556v"] Oct 02 16:55:57 crc kubenswrapper[4808]: W1002 16:55:57.428208 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod181f7b50_6e00_4e8c_bdde_9bbff23c3306.slice/crio-d739ccd5aa66f1f2bbe82694fdbe9ed20d90d094dd768e78812394d483779901 WatchSource:0}: Error finding container d739ccd5aa66f1f2bbe82694fdbe9ed20d90d094dd768e78812394d483779901: Status 404 returned error can't find the container with id d739ccd5aa66f1f2bbe82694fdbe9ed20d90d094dd768e78812394d483779901 Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.552973 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f18f563e-79e6-4c33-87a5-91f7c27bf181","Type":"ContainerStarted","Data":"b6e973576609e6ccbd8861e21134aebcda316cfd3aa63914174cc4babd323e49"} Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.553976 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-l556v" event={"ID":"181f7b50-6e00-4e8c-bdde-9bbff23c3306","Type":"ContainerStarted","Data":"d739ccd5aa66f1f2bbe82694fdbe9ed20d90d094dd768e78812394d483779901"} Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.555411 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hzbv4" event={"ID":"eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8","Type":"ContainerStarted","Data":"2c9821b4b2844d0ca67a5d994ef76918a735576ede1d192cee41976f9b430903"} Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.555636 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-hzbv4" Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.558535 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b409b3d8-0ed6-49e9-a1ea-be6689e607ed","Type":"ContainerStarted","Data":"40f7dc7f29d9aa9488c1ee47ca18e78ac062b2d82976a34fcdbda5c0ee3df075"} Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.560789 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"500b4d21-bf0e-4763-91bd-7c32bf13a064","Type":"ContainerStarted","Data":"a80100e6984b9ad4b05d503e868828d74a33b70920d7244bb93b8882f0694875"} Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.562325 4808 generic.go:334] "Generic (PLEG): container finished" podID="fd2a1d38-1169-4ab7-995f-eb6fb2f230a6" containerID="55ecdbee6e01f02df2b7c842500b8c7ac2b1222ad64cd246def2328b5ddf41e7" exitCode=0 Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.562385 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-59c8c" event={"ID":"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6","Type":"ContainerDied","Data":"55ecdbee6e01f02df2b7c842500b8c7ac2b1222ad64cd246def2328b5ddf41e7"} Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.570643 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f71818d6-1ebf-4838-b871-8873c107b325","Type":"ContainerStarted","Data":"3733dae6ab12fcdb987673b7dd9fdcbf7a55150a8de5aa8be5de065407e9afcf"} Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.573359 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-qgt5h" Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.573746 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-8d9tc" event={"ID":"ebecde28-66be-4caa-87d4-779d8605da87","Type":"ContainerStarted","Data":"a8cc68871a8693e11330ac5365a276ea38af84f53607ee5d9bf34f7d26e7fbd9"} Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.573840 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-hzbv4" podStartSLOduration=5.382866013 podStartE2EDuration="13.573823837s" podCreationTimestamp="2025-10-02 16:55:44 +0000 UTC" firstStartedPulling="2025-10-02 16:55:47.134972845 +0000 UTC m=+914.460501845" lastFinishedPulling="2025-10-02 16:55:55.325930659 +0000 UTC m=+922.651459669" observedRunningTime="2025-10-02 16:55:57.570708013 +0000 UTC m=+924.896237013" watchObservedRunningTime="2025-10-02 16:55:57.573823837 +0000 UTC m=+924.899352837" Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.585354 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-qgt5h" Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.699582 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d38d552b-cc48-4ad2-b6df-873129d015e8-config\") pod \"d38d552b-cc48-4ad2-b6df-873129d015e8\" (UID: \"d38d552b-cc48-4ad2-b6df-873129d015e8\") " Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.699657 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m24lp\" (UniqueName: \"kubernetes.io/projected/d38d552b-cc48-4ad2-b6df-873129d015e8-kube-api-access-m24lp\") pod \"d38d552b-cc48-4ad2-b6df-873129d015e8\" (UID: \"d38d552b-cc48-4ad2-b6df-873129d015e8\") " Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.699776 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d38d552b-cc48-4ad2-b6df-873129d015e8-ovsdbserver-nb\") pod \"d38d552b-cc48-4ad2-b6df-873129d015e8\" (UID: \"d38d552b-cc48-4ad2-b6df-873129d015e8\") " Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.699866 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d38d552b-cc48-4ad2-b6df-873129d015e8-dns-svc\") pod \"d38d552b-cc48-4ad2-b6df-873129d015e8\" (UID: \"d38d552b-cc48-4ad2-b6df-873129d015e8\") " Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.700417 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d38d552b-cc48-4ad2-b6df-873129d015e8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d38d552b-cc48-4ad2-b6df-873129d015e8" (UID: "d38d552b-cc48-4ad2-b6df-873129d015e8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.701122 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d38d552b-cc48-4ad2-b6df-873129d015e8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.701176 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d38d552b-cc48-4ad2-b6df-873129d015e8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d38d552b-cc48-4ad2-b6df-873129d015e8" (UID: "d38d552b-cc48-4ad2-b6df-873129d015e8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.701425 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d38d552b-cc48-4ad2-b6df-873129d015e8-config" (OuterVolumeSpecName: "config") pod "d38d552b-cc48-4ad2-b6df-873129d015e8" (UID: "d38d552b-cc48-4ad2-b6df-873129d015e8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.708611 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d38d552b-cc48-4ad2-b6df-873129d015e8-kube-api-access-m24lp" (OuterVolumeSpecName: "kube-api-access-m24lp") pod "d38d552b-cc48-4ad2-b6df-873129d015e8" (UID: "d38d552b-cc48-4ad2-b6df-873129d015e8"). InnerVolumeSpecName "kube-api-access-m24lp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.806747 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d38d552b-cc48-4ad2-b6df-873129d015e8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.806777 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d38d552b-cc48-4ad2-b6df-873129d015e8-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:55:57 crc kubenswrapper[4808]: I1002 16:55:57.806790 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m24lp\" (UniqueName: \"kubernetes.io/projected/d38d552b-cc48-4ad2-b6df-873129d015e8-kube-api-access-m24lp\") on node \"crc\" DevicePath \"\"" Oct 02 16:55:58 crc kubenswrapper[4808]: I1002 16:55:58.585396 4808 generic.go:334] "Generic (PLEG): container finished" podID="181f7b50-6e00-4e8c-bdde-9bbff23c3306" containerID="f662f26ed8c912737d56a4c4f68e273cd086711a4650a7ae755f89b9883b82be" exitCode=0 Oct 02 16:55:58 crc kubenswrapper[4808]: I1002 16:55:58.585540 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-l556v" event={"ID":"181f7b50-6e00-4e8c-bdde-9bbff23c3306","Type":"ContainerDied","Data":"f662f26ed8c912737d56a4c4f68e273cd086711a4650a7ae755f89b9883b82be"} Oct 02 16:55:58 crc kubenswrapper[4808]: I1002 16:55:58.589707 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-59c8c" event={"ID":"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6","Type":"ContainerStarted","Data":"618368d22f49959531728d4eb9e19055e75dc6c259c28fdf70637fd94e646265"} Oct 02 16:55:58 crc kubenswrapper[4808]: I1002 16:55:58.589761 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-59c8c" event={"ID":"fd2a1d38-1169-4ab7-995f-eb6fb2f230a6","Type":"ContainerStarted","Data":"bb2004eb68a90f6b1426c83badef5957442f85702aaed522b4d5ca2ee7691157"} Oct 02 16:55:58 crc kubenswrapper[4808]: I1002 16:55:58.590288 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-qgt5h" Oct 02 16:55:58 crc kubenswrapper[4808]: I1002 16:55:58.649408 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-qgt5h"] Oct 02 16:55:58 crc kubenswrapper[4808]: I1002 16:55:58.654202 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-qgt5h"] Oct 02 16:55:58 crc kubenswrapper[4808]: I1002 16:55:58.666899 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-59c8c" podStartSLOduration=6.81749777 podStartE2EDuration="14.666883273s" podCreationTimestamp="2025-10-02 16:55:44 +0000 UTC" firstStartedPulling="2025-10-02 16:55:47.476527056 +0000 UTC m=+914.802056056" lastFinishedPulling="2025-10-02 16:55:55.325912559 +0000 UTC m=+922.651441559" observedRunningTime="2025-10-02 16:55:58.665037783 +0000 UTC m=+925.990566793" watchObservedRunningTime="2025-10-02 16:55:58.666883273 +0000 UTC m=+925.992412273" Oct 02 16:55:59 crc kubenswrapper[4808]: I1002 16:55:59.416394 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d38d552b-cc48-4ad2-b6df-873129d015e8" path="/var/lib/kubelet/pods/d38d552b-cc48-4ad2-b6df-873129d015e8/volumes" Oct 02 16:55:59 crc kubenswrapper[4808]: I1002 16:55:59.872930 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:55:59 crc kubenswrapper[4808]: I1002 16:55:59.873005 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:56:01 crc kubenswrapper[4808]: I1002 16:56:01.632878 4808 generic.go:334] "Generic (PLEG): container finished" podID="500b4d21-bf0e-4763-91bd-7c32bf13a064" containerID="a80100e6984b9ad4b05d503e868828d74a33b70920d7244bb93b8882f0694875" exitCode=0 Oct 02 16:56:01 crc kubenswrapper[4808]: I1002 16:56:01.633030 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"500b4d21-bf0e-4763-91bd-7c32bf13a064","Type":"ContainerDied","Data":"a80100e6984b9ad4b05d503e868828d74a33b70920d7244bb93b8882f0694875"} Oct 02 16:56:01 crc kubenswrapper[4808]: I1002 16:56:01.636737 4808 generic.go:334] "Generic (PLEG): container finished" podID="a8509820-f0d7-417b-92d4-1be8b7b714ef" containerID="fe6b573c8f477b53c3c392d96d8d584666a437a31a77b0a91bc86d96a6fc55c0" exitCode=0 Oct 02 16:56:01 crc kubenswrapper[4808]: I1002 16:56:01.636819 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a8509820-f0d7-417b-92d4-1be8b7b714ef","Type":"ContainerDied","Data":"fe6b573c8f477b53c3c392d96d8d584666a437a31a77b0a91bc86d96a6fc55c0"} Oct 02 16:56:03 crc kubenswrapper[4808]: I1002 16:56:03.647519 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 02 16:56:08 crc kubenswrapper[4808]: I1002 16:56:08.711988 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"dbb01d1d-5fc2-4d1c-8fa3-5575de879622","Type":"ContainerStarted","Data":"01277088a416a9f21f6825370c6729633ba695fb99654aa70ff8cac39056fb33"} Oct 02 16:56:08 crc kubenswrapper[4808]: I1002 16:56:08.714788 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d642175e-72ec-466b-9273-b82beb5e99f5","Type":"ContainerStarted","Data":"d47be86869ed0a5926a51529c11ed74292ac95c72dff27e0d788e94764a7569c"} Oct 02 16:56:08 crc kubenswrapper[4808]: I1002 16:56:08.717916 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"500b4d21-bf0e-4763-91bd-7c32bf13a064","Type":"ContainerStarted","Data":"033001578c059ea03011911077a273625ece487924a4b1813c2b7b6bb3965450"} Oct 02 16:56:08 crc kubenswrapper[4808]: I1002 16:56:08.720570 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-8d9tc" event={"ID":"ebecde28-66be-4caa-87d4-779d8605da87","Type":"ContainerStarted","Data":"c26adc3f773c8f30259900d279dda8d43be41f85e072c9c361e39fa88cd25b5e"} Oct 02 16:56:08 crc kubenswrapper[4808]: I1002 16:56:08.723620 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a8509820-f0d7-417b-92d4-1be8b7b714ef","Type":"ContainerStarted","Data":"6cf64a163c6fe52f68a4c2e7f455dd7e34d451ea410fbed4633304dcb11c73a6"} Oct 02 16:56:08 crc kubenswrapper[4808]: I1002 16:56:08.726999 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-l556v" event={"ID":"181f7b50-6e00-4e8c-bdde-9bbff23c3306","Type":"ContainerStarted","Data":"23c860288b64583e7ca3a50a120dcfd0ae8d8d7269348d26f8c69828f1139196"} Oct 02 16:56:08 crc kubenswrapper[4808]: I1002 16:56:08.727199 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:56:08 crc kubenswrapper[4808]: I1002 16:56:08.745651 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.807359793 podStartE2EDuration="23.745624691s" podCreationTimestamp="2025-10-02 16:55:45 +0000 UTC" firstStartedPulling="2025-10-02 16:55:47.934637545 +0000 UTC m=+915.260166545" lastFinishedPulling="2025-10-02 16:56:07.872902443 +0000 UTC m=+935.198431443" observedRunningTime="2025-10-02 16:56:08.744611763 +0000 UTC m=+936.070140803" watchObservedRunningTime="2025-10-02 16:56:08.745624691 +0000 UTC m=+936.071153731" Oct 02 16:56:08 crc kubenswrapper[4808]: I1002 16:56:08.793537 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=24.509880153 podStartE2EDuration="32.793502218s" podCreationTimestamp="2025-10-02 16:55:36 +0000 UTC" firstStartedPulling="2025-10-02 16:55:47.206100131 +0000 UTC m=+914.531629141" lastFinishedPulling="2025-10-02 16:55:55.489722206 +0000 UTC m=+922.815251206" observedRunningTime="2025-10-02 16:56:08.781871413 +0000 UTC m=+936.107400503" watchObservedRunningTime="2025-10-02 16:56:08.793502218 +0000 UTC m=+936.119031258" Oct 02 16:56:08 crc kubenswrapper[4808]: I1002 16:56:08.828850 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.361901986 podStartE2EDuration="24.828824764s" podCreationTimestamp="2025-10-02 16:55:44 +0000 UTC" firstStartedPulling="2025-10-02 16:55:47.35891887 +0000 UTC m=+914.684447870" lastFinishedPulling="2025-10-02 16:56:07.825841648 +0000 UTC m=+935.151370648" observedRunningTime="2025-10-02 16:56:08.821415074 +0000 UTC m=+936.146944154" watchObservedRunningTime="2025-10-02 16:56:08.828824764 +0000 UTC m=+936.154353774" Oct 02 16:56:08 crc kubenswrapper[4808]: I1002 16:56:08.859889 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-8d9tc" podStartSLOduration=2.098986281 podStartE2EDuration="12.859858935s" podCreationTimestamp="2025-10-02 16:55:56 +0000 UTC" firstStartedPulling="2025-10-02 16:55:57.112165112 +0000 UTC m=+924.437694102" lastFinishedPulling="2025-10-02 16:56:07.873037716 +0000 UTC m=+935.198566756" observedRunningTime="2025-10-02 16:56:08.850951744 +0000 UTC m=+936.176480764" watchObservedRunningTime="2025-10-02 16:56:08.859858935 +0000 UTC m=+936.185387975" Oct 02 16:56:08 crc kubenswrapper[4808]: I1002 16:56:08.878031 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-l556v" podStartSLOduration=12.878012157 podStartE2EDuration="12.878012157s" podCreationTimestamp="2025-10-02 16:55:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:56:08.87481153 +0000 UTC m=+936.200340540" watchObservedRunningTime="2025-10-02 16:56:08.878012157 +0000 UTC m=+936.203541167" Oct 02 16:56:08 crc kubenswrapper[4808]: I1002 16:56:08.915694 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=24.188404775 podStartE2EDuration="32.915673157s" podCreationTimestamp="2025-10-02 16:55:36 +0000 UTC" firstStartedPulling="2025-10-02 16:55:47.228135348 +0000 UTC m=+914.553664358" lastFinishedPulling="2025-10-02 16:55:55.95540373 +0000 UTC m=+923.280932740" observedRunningTime="2025-10-02 16:56:08.910130637 +0000 UTC m=+936.235659657" watchObservedRunningTime="2025-10-02 16:56:08.915673157 +0000 UTC m=+936.241202167" Oct 02 16:56:09 crc kubenswrapper[4808]: I1002 16:56:09.476268 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 02 16:56:09 crc kubenswrapper[4808]: I1002 16:56:09.540049 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 02 16:56:09 crc kubenswrapper[4808]: I1002 16:56:09.736968 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 02 16:56:09 crc kubenswrapper[4808]: I1002 16:56:09.799843 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 02 16:56:10 crc kubenswrapper[4808]: I1002 16:56:10.477290 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 16:56:11 crc kubenswrapper[4808]: I1002 16:56:11.264806 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 02 16:56:11 crc kubenswrapper[4808]: I1002 16:56:11.338496 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 02 16:56:11 crc kubenswrapper[4808]: I1002 16:56:11.751762 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 02 16:56:11 crc kubenswrapper[4808]: I1002 16:56:11.823159 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.014849 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.016063 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.022068 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.022079 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.022885 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.022973 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-68s45" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.034278 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.164526 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvdwh\" (UniqueName: \"kubernetes.io/projected/c84f87a5-f6b6-49b9-aa85-e88695deaa87-kube-api-access-bvdwh\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.164563 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c84f87a5-f6b6-49b9-aa85-e88695deaa87-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.164593 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c84f87a5-f6b6-49b9-aa85-e88695deaa87-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.164617 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c84f87a5-f6b6-49b9-aa85-e88695deaa87-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.164637 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c84f87a5-f6b6-49b9-aa85-e88695deaa87-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.164655 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c84f87a5-f6b6-49b9-aa85-e88695deaa87-config\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.164673 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c84f87a5-f6b6-49b9-aa85-e88695deaa87-scripts\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.265582 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvdwh\" (UniqueName: \"kubernetes.io/projected/c84f87a5-f6b6-49b9-aa85-e88695deaa87-kube-api-access-bvdwh\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.265632 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c84f87a5-f6b6-49b9-aa85-e88695deaa87-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.265664 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c84f87a5-f6b6-49b9-aa85-e88695deaa87-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.265690 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c84f87a5-f6b6-49b9-aa85-e88695deaa87-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.265707 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c84f87a5-f6b6-49b9-aa85-e88695deaa87-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.265725 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c84f87a5-f6b6-49b9-aa85-e88695deaa87-config\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.265741 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c84f87a5-f6b6-49b9-aa85-e88695deaa87-scripts\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.266861 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c84f87a5-f6b6-49b9-aa85-e88695deaa87-scripts\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.267484 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c84f87a5-f6b6-49b9-aa85-e88695deaa87-config\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.269555 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c84f87a5-f6b6-49b9-aa85-e88695deaa87-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.271982 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c84f87a5-f6b6-49b9-aa85-e88695deaa87-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.283339 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvdwh\" (UniqueName: \"kubernetes.io/projected/c84f87a5-f6b6-49b9-aa85-e88695deaa87-kube-api-access-bvdwh\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.286605 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c84f87a5-f6b6-49b9-aa85-e88695deaa87-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.287807 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c84f87a5-f6b6-49b9-aa85-e88695deaa87-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c84f87a5-f6b6-49b9-aa85-e88695deaa87\") " pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.357809 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 16:56:12 crc kubenswrapper[4808]: I1002 16:56:12.870389 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 16:56:12 crc kubenswrapper[4808]: W1002 16:56:12.879498 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc84f87a5_f6b6_49b9_aa85_e88695deaa87.slice/crio-e805cd294622880c7c25bb96f27516412cdc37e450d91a30006387ffec7230e6 WatchSource:0}: Error finding container e805cd294622880c7c25bb96f27516412cdc37e450d91a30006387ffec7230e6: Status 404 returned error can't find the container with id e805cd294622880c7c25bb96f27516412cdc37e450d91a30006387ffec7230e6 Oct 02 16:56:13 crc kubenswrapper[4808]: I1002 16:56:13.769996 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c84f87a5-f6b6-49b9-aa85-e88695deaa87","Type":"ContainerStarted","Data":"e805cd294622880c7c25bb96f27516412cdc37e450d91a30006387ffec7230e6"} Oct 02 16:56:14 crc kubenswrapper[4808]: I1002 16:56:14.783571 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c84f87a5-f6b6-49b9-aa85-e88695deaa87","Type":"ContainerStarted","Data":"5193fbea78f6eb5176f1f735c34c2a85b71a796d7b7532cf7d1de431040f3784"} Oct 02 16:56:14 crc kubenswrapper[4808]: I1002 16:56:14.784208 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 02 16:56:14 crc kubenswrapper[4808]: I1002 16:56:14.784230 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c84f87a5-f6b6-49b9-aa85-e88695deaa87","Type":"ContainerStarted","Data":"927dcfc2d9a69ebbd479e18d9f5c12f344a73fc7b9158683d3717cce4e26638e"} Oct 02 16:56:14 crc kubenswrapper[4808]: I1002 16:56:14.818058 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.6333353710000003 podStartE2EDuration="3.818034911s" podCreationTimestamp="2025-10-02 16:56:11 +0000 UTC" firstStartedPulling="2025-10-02 16:56:12.885355991 +0000 UTC m=+940.210885011" lastFinishedPulling="2025-10-02 16:56:14.070055541 +0000 UTC m=+941.395584551" observedRunningTime="2025-10-02 16:56:14.808524513 +0000 UTC m=+942.134053593" watchObservedRunningTime="2025-10-02 16:56:14.818034911 +0000 UTC m=+942.143563921" Oct 02 16:56:16 crc kubenswrapper[4808]: I1002 16:56:16.986457 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.066524 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-xgpxx"] Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.069670 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" podUID="45c90d4a-2e61-431a-97b4-900c0ceb6493" containerName="dnsmasq-dns" containerID="cri-o://7a105c68d069497311e69ee8c3f8111d30a9957c091c2b271a728e67b7239eba" gracePeriod=10 Oct 02 16:56:17 crc kubenswrapper[4808]: E1002 16:56:17.215251 4808 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45c90d4a_2e61_431a_97b4_900c0ceb6493.slice/crio-7a105c68d069497311e69ee8c3f8111d30a9957c091c2b271a728e67b7239eba.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45c90d4a_2e61_431a_97b4_900c0ceb6493.slice/crio-conmon-7a105c68d069497311e69ee8c3f8111d30a9957c091c2b271a728e67b7239eba.scope\": RecentStats: unable to find data in memory cache]" Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.589830 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.667555 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brfn7\" (UniqueName: \"kubernetes.io/projected/45c90d4a-2e61-431a-97b4-900c0ceb6493-kube-api-access-brfn7\") pod \"45c90d4a-2e61-431a-97b4-900c0ceb6493\" (UID: \"45c90d4a-2e61-431a-97b4-900c0ceb6493\") " Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.667635 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45c90d4a-2e61-431a-97b4-900c0ceb6493-dns-svc\") pod \"45c90d4a-2e61-431a-97b4-900c0ceb6493\" (UID: \"45c90d4a-2e61-431a-97b4-900c0ceb6493\") " Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.667732 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45c90d4a-2e61-431a-97b4-900c0ceb6493-config\") pod \"45c90d4a-2e61-431a-97b4-900c0ceb6493\" (UID: \"45c90d4a-2e61-431a-97b4-900c0ceb6493\") " Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.682463 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45c90d4a-2e61-431a-97b4-900c0ceb6493-kube-api-access-brfn7" (OuterVolumeSpecName: "kube-api-access-brfn7") pod "45c90d4a-2e61-431a-97b4-900c0ceb6493" (UID: "45c90d4a-2e61-431a-97b4-900c0ceb6493"). InnerVolumeSpecName "kube-api-access-brfn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.709054 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45c90d4a-2e61-431a-97b4-900c0ceb6493-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "45c90d4a-2e61-431a-97b4-900c0ceb6493" (UID: "45c90d4a-2e61-431a-97b4-900c0ceb6493"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.736889 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45c90d4a-2e61-431a-97b4-900c0ceb6493-config" (OuterVolumeSpecName: "config") pod "45c90d4a-2e61-431a-97b4-900c0ceb6493" (UID: "45c90d4a-2e61-431a-97b4-900c0ceb6493"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.769180 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brfn7\" (UniqueName: \"kubernetes.io/projected/45c90d4a-2e61-431a-97b4-900c0ceb6493-kube-api-access-brfn7\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.769209 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45c90d4a-2e61-431a-97b4-900c0ceb6493-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.769219 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45c90d4a-2e61-431a-97b4-900c0ceb6493-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.811227 4808 generic.go:334] "Generic (PLEG): container finished" podID="45c90d4a-2e61-431a-97b4-900c0ceb6493" containerID="7a105c68d069497311e69ee8c3f8111d30a9957c091c2b271a728e67b7239eba" exitCode=0 Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.811287 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" event={"ID":"45c90d4a-2e61-431a-97b4-900c0ceb6493","Type":"ContainerDied","Data":"7a105c68d069497311e69ee8c3f8111d30a9957c091c2b271a728e67b7239eba"} Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.811311 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" event={"ID":"45c90d4a-2e61-431a-97b4-900c0ceb6493","Type":"ContainerDied","Data":"fac50baba4bb7f5df8fb1f22ab82167eb2d30460de34191414022f896e0c8b6d"} Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.811313 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-xgpxx" Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.811328 4808 scope.go:117] "RemoveContainer" containerID="7a105c68d069497311e69ee8c3f8111d30a9957c091c2b271a728e67b7239eba" Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.833781 4808 scope.go:117] "RemoveContainer" containerID="6f3066add9aae146979e59164ea6c6224346e492c34c1772e8493e7651d78036" Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.852289 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-xgpxx"] Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.871689 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-xgpxx"] Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.876049 4808 scope.go:117] "RemoveContainer" containerID="7a105c68d069497311e69ee8c3f8111d30a9957c091c2b271a728e67b7239eba" Oct 02 16:56:17 crc kubenswrapper[4808]: E1002 16:56:17.876581 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a105c68d069497311e69ee8c3f8111d30a9957c091c2b271a728e67b7239eba\": container with ID starting with 7a105c68d069497311e69ee8c3f8111d30a9957c091c2b271a728e67b7239eba not found: ID does not exist" containerID="7a105c68d069497311e69ee8c3f8111d30a9957c091c2b271a728e67b7239eba" Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.876639 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a105c68d069497311e69ee8c3f8111d30a9957c091c2b271a728e67b7239eba"} err="failed to get container status \"7a105c68d069497311e69ee8c3f8111d30a9957c091c2b271a728e67b7239eba\": rpc error: code = NotFound desc = could not find container \"7a105c68d069497311e69ee8c3f8111d30a9957c091c2b271a728e67b7239eba\": container with ID starting with 7a105c68d069497311e69ee8c3f8111d30a9957c091c2b271a728e67b7239eba not found: ID does not exist" Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.876673 4808 scope.go:117] "RemoveContainer" containerID="6f3066add9aae146979e59164ea6c6224346e492c34c1772e8493e7651d78036" Oct 02 16:56:17 crc kubenswrapper[4808]: E1002 16:56:17.877118 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f3066add9aae146979e59164ea6c6224346e492c34c1772e8493e7651d78036\": container with ID starting with 6f3066add9aae146979e59164ea6c6224346e492c34c1772e8493e7651d78036 not found: ID does not exist" containerID="6f3066add9aae146979e59164ea6c6224346e492c34c1772e8493e7651d78036" Oct 02 16:56:17 crc kubenswrapper[4808]: I1002 16:56:17.877147 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f3066add9aae146979e59164ea6c6224346e492c34c1772e8493e7651d78036"} err="failed to get container status \"6f3066add9aae146979e59164ea6c6224346e492c34c1772e8493e7651d78036\": rpc error: code = NotFound desc = could not find container \"6f3066add9aae146979e59164ea6c6224346e492c34c1772e8493e7651d78036\": container with ID starting with 6f3066add9aae146979e59164ea6c6224346e492c34c1772e8493e7651d78036 not found: ID does not exist" Oct 02 16:56:18 crc kubenswrapper[4808]: I1002 16:56:18.158889 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 02 16:56:18 crc kubenswrapper[4808]: I1002 16:56:18.162657 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 02 16:56:18 crc kubenswrapper[4808]: I1002 16:56:18.228099 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 02 16:56:18 crc kubenswrapper[4808]: I1002 16:56:18.241927 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 02 16:56:18 crc kubenswrapper[4808]: I1002 16:56:18.242287 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 02 16:56:18 crc kubenswrapper[4808]: I1002 16:56:18.303194 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 02 16:56:18 crc kubenswrapper[4808]: I1002 16:56:18.887938 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 02 16:56:18 crc kubenswrapper[4808]: I1002 16:56:18.898669 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 02 16:56:19 crc kubenswrapper[4808]: I1002 16:56:19.410518 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45c90d4a-2e61-431a-97b4-900c0ceb6493" path="/var/lib/kubelet/pods/45c90d4a-2e61-431a-97b4-900c0ceb6493/volumes" Oct 02 16:56:23 crc kubenswrapper[4808]: I1002 16:56:23.967574 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-rk9zb"] Oct 02 16:56:23 crc kubenswrapper[4808]: E1002 16:56:23.968403 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45c90d4a-2e61-431a-97b4-900c0ceb6493" containerName="init" Oct 02 16:56:23 crc kubenswrapper[4808]: I1002 16:56:23.968427 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="45c90d4a-2e61-431a-97b4-900c0ceb6493" containerName="init" Oct 02 16:56:23 crc kubenswrapper[4808]: E1002 16:56:23.968459 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45c90d4a-2e61-431a-97b4-900c0ceb6493" containerName="dnsmasq-dns" Oct 02 16:56:23 crc kubenswrapper[4808]: I1002 16:56:23.968473 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="45c90d4a-2e61-431a-97b4-900c0ceb6493" containerName="dnsmasq-dns" Oct 02 16:56:23 crc kubenswrapper[4808]: I1002 16:56:23.968761 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="45c90d4a-2e61-431a-97b4-900c0ceb6493" containerName="dnsmasq-dns" Oct 02 16:56:23 crc kubenswrapper[4808]: I1002 16:56:23.969630 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rk9zb" Oct 02 16:56:23 crc kubenswrapper[4808]: I1002 16:56:23.980303 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-rk9zb"] Oct 02 16:56:23 crc kubenswrapper[4808]: I1002 16:56:23.989668 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp9lv\" (UniqueName: \"kubernetes.io/projected/16cb8757-fcb8-469a-8971-1c6dffa2e3fd-kube-api-access-tp9lv\") pod \"glance-db-create-rk9zb\" (UID: \"16cb8757-fcb8-469a-8971-1c6dffa2e3fd\") " pod="openstack/glance-db-create-rk9zb" Oct 02 16:56:24 crc kubenswrapper[4808]: I1002 16:56:24.091420 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp9lv\" (UniqueName: \"kubernetes.io/projected/16cb8757-fcb8-469a-8971-1c6dffa2e3fd-kube-api-access-tp9lv\") pod \"glance-db-create-rk9zb\" (UID: \"16cb8757-fcb8-469a-8971-1c6dffa2e3fd\") " pod="openstack/glance-db-create-rk9zb" Oct 02 16:56:24 crc kubenswrapper[4808]: I1002 16:56:24.115701 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp9lv\" (UniqueName: \"kubernetes.io/projected/16cb8757-fcb8-469a-8971-1c6dffa2e3fd-kube-api-access-tp9lv\") pod \"glance-db-create-rk9zb\" (UID: \"16cb8757-fcb8-469a-8971-1c6dffa2e3fd\") " pod="openstack/glance-db-create-rk9zb" Oct 02 16:56:24 crc kubenswrapper[4808]: I1002 16:56:24.291558 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rk9zb" Oct 02 16:56:24 crc kubenswrapper[4808]: I1002 16:56:24.775456 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-rk9zb"] Oct 02 16:56:24 crc kubenswrapper[4808]: W1002 16:56:24.779990 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16cb8757_fcb8_469a_8971_1c6dffa2e3fd.slice/crio-e2441bc7a2540ee85131f421f318285b183d2eea89e10ebe917ebc3dc01045d8 WatchSource:0}: Error finding container e2441bc7a2540ee85131f421f318285b183d2eea89e10ebe917ebc3dc01045d8: Status 404 returned error can't find the container with id e2441bc7a2540ee85131f421f318285b183d2eea89e10ebe917ebc3dc01045d8 Oct 02 16:56:24 crc kubenswrapper[4808]: I1002 16:56:24.886698 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rk9zb" event={"ID":"16cb8757-fcb8-469a-8971-1c6dffa2e3fd","Type":"ContainerStarted","Data":"e2441bc7a2540ee85131f421f318285b183d2eea89e10ebe917ebc3dc01045d8"} Oct 02 16:56:25 crc kubenswrapper[4808]: I1002 16:56:25.906145 4808 generic.go:334] "Generic (PLEG): container finished" podID="16cb8757-fcb8-469a-8971-1c6dffa2e3fd" containerID="7540bac8b871f290c1c2c2b8ce6c7ebbeace8f4dddb12fe60a0a4225b17e3b29" exitCode=0 Oct 02 16:56:25 crc kubenswrapper[4808]: I1002 16:56:25.906260 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rk9zb" event={"ID":"16cb8757-fcb8-469a-8971-1c6dffa2e3fd","Type":"ContainerDied","Data":"7540bac8b871f290c1c2c2b8ce6c7ebbeace8f4dddb12fe60a0a4225b17e3b29"} Oct 02 16:56:27 crc kubenswrapper[4808]: I1002 16:56:27.336987 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rk9zb" Oct 02 16:56:27 crc kubenswrapper[4808]: I1002 16:56:27.369370 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tp9lv\" (UniqueName: \"kubernetes.io/projected/16cb8757-fcb8-469a-8971-1c6dffa2e3fd-kube-api-access-tp9lv\") pod \"16cb8757-fcb8-469a-8971-1c6dffa2e3fd\" (UID: \"16cb8757-fcb8-469a-8971-1c6dffa2e3fd\") " Oct 02 16:56:27 crc kubenswrapper[4808]: I1002 16:56:27.382334 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16cb8757-fcb8-469a-8971-1c6dffa2e3fd-kube-api-access-tp9lv" (OuterVolumeSpecName: "kube-api-access-tp9lv") pod "16cb8757-fcb8-469a-8971-1c6dffa2e3fd" (UID: "16cb8757-fcb8-469a-8971-1c6dffa2e3fd"). InnerVolumeSpecName "kube-api-access-tp9lv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:56:27 crc kubenswrapper[4808]: I1002 16:56:27.444446 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 02 16:56:27 crc kubenswrapper[4808]: I1002 16:56:27.473280 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tp9lv\" (UniqueName: \"kubernetes.io/projected/16cb8757-fcb8-469a-8971-1c6dffa2e3fd-kube-api-access-tp9lv\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:27 crc kubenswrapper[4808]: I1002 16:56:27.936201 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rk9zb" event={"ID":"16cb8757-fcb8-469a-8971-1c6dffa2e3fd","Type":"ContainerDied","Data":"e2441bc7a2540ee85131f421f318285b183d2eea89e10ebe917ebc3dc01045d8"} Oct 02 16:56:27 crc kubenswrapper[4808]: I1002 16:56:27.937120 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2441bc7a2540ee85131f421f318285b183d2eea89e10ebe917ebc3dc01045d8" Oct 02 16:56:27 crc kubenswrapper[4808]: I1002 16:56:27.936367 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rk9zb" Oct 02 16:56:28 crc kubenswrapper[4808]: I1002 16:56:28.281145 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-twjnb"] Oct 02 16:56:28 crc kubenswrapper[4808]: E1002 16:56:28.281924 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16cb8757-fcb8-469a-8971-1c6dffa2e3fd" containerName="mariadb-database-create" Oct 02 16:56:28 crc kubenswrapper[4808]: I1002 16:56:28.281956 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="16cb8757-fcb8-469a-8971-1c6dffa2e3fd" containerName="mariadb-database-create" Oct 02 16:56:28 crc kubenswrapper[4808]: I1002 16:56:28.282427 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="16cb8757-fcb8-469a-8971-1c6dffa2e3fd" containerName="mariadb-database-create" Oct 02 16:56:28 crc kubenswrapper[4808]: I1002 16:56:28.283514 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-twjnb" Oct 02 16:56:28 crc kubenswrapper[4808]: I1002 16:56:28.290649 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-twjnb"] Oct 02 16:56:28 crc kubenswrapper[4808]: I1002 16:56:28.389224 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wpbc\" (UniqueName: \"kubernetes.io/projected/06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c-kube-api-access-6wpbc\") pod \"keystone-db-create-twjnb\" (UID: \"06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c\") " pod="openstack/keystone-db-create-twjnb" Oct 02 16:56:28 crc kubenswrapper[4808]: I1002 16:56:28.493071 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wpbc\" (UniqueName: \"kubernetes.io/projected/06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c-kube-api-access-6wpbc\") pod \"keystone-db-create-twjnb\" (UID: \"06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c\") " pod="openstack/keystone-db-create-twjnb" Oct 02 16:56:28 crc kubenswrapper[4808]: I1002 16:56:28.525915 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wpbc\" (UniqueName: \"kubernetes.io/projected/06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c-kube-api-access-6wpbc\") pod \"keystone-db-create-twjnb\" (UID: \"06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c\") " pod="openstack/keystone-db-create-twjnb" Oct 02 16:56:28 crc kubenswrapper[4808]: I1002 16:56:28.583894 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-5fthx"] Oct 02 16:56:28 crc kubenswrapper[4808]: I1002 16:56:28.585728 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5fthx" Oct 02 16:56:28 crc kubenswrapper[4808]: I1002 16:56:28.594078 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-5fthx"] Oct 02 16:56:28 crc kubenswrapper[4808]: I1002 16:56:28.601449 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-twjnb" Oct 02 16:56:28 crc kubenswrapper[4808]: I1002 16:56:28.695985 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htxbm\" (UniqueName: \"kubernetes.io/projected/0b6c4102-3ccd-4119-bbad-697ad2589e40-kube-api-access-htxbm\") pod \"placement-db-create-5fthx\" (UID: \"0b6c4102-3ccd-4119-bbad-697ad2589e40\") " pod="openstack/placement-db-create-5fthx" Oct 02 16:56:28 crc kubenswrapper[4808]: I1002 16:56:28.798190 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htxbm\" (UniqueName: \"kubernetes.io/projected/0b6c4102-3ccd-4119-bbad-697ad2589e40-kube-api-access-htxbm\") pod \"placement-db-create-5fthx\" (UID: \"0b6c4102-3ccd-4119-bbad-697ad2589e40\") " pod="openstack/placement-db-create-5fthx" Oct 02 16:56:28 crc kubenswrapper[4808]: I1002 16:56:28.838343 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htxbm\" (UniqueName: \"kubernetes.io/projected/0b6c4102-3ccd-4119-bbad-697ad2589e40-kube-api-access-htxbm\") pod \"placement-db-create-5fthx\" (UID: \"0b6c4102-3ccd-4119-bbad-697ad2589e40\") " pod="openstack/placement-db-create-5fthx" Oct 02 16:56:28 crc kubenswrapper[4808]: I1002 16:56:28.923534 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5fthx" Oct 02 16:56:28 crc kubenswrapper[4808]: I1002 16:56:28.946011 4808 generic.go:334] "Generic (PLEG): container finished" podID="f71818d6-1ebf-4838-b871-8873c107b325" containerID="3733dae6ab12fcdb987673b7dd9fdcbf7a55150a8de5aa8be5de065407e9afcf" exitCode=0 Oct 02 16:56:28 crc kubenswrapper[4808]: I1002 16:56:28.946085 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f71818d6-1ebf-4838-b871-8873c107b325","Type":"ContainerDied","Data":"3733dae6ab12fcdb987673b7dd9fdcbf7a55150a8de5aa8be5de065407e9afcf"} Oct 02 16:56:28 crc kubenswrapper[4808]: I1002 16:56:28.948190 4808 generic.go:334] "Generic (PLEG): container finished" podID="b409b3d8-0ed6-49e9-a1ea-be6689e607ed" containerID="40f7dc7f29d9aa9488c1ee47ca18e78ac062b2d82976a34fcdbda5c0ee3df075" exitCode=0 Oct 02 16:56:28 crc kubenswrapper[4808]: I1002 16:56:28.948282 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b409b3d8-0ed6-49e9-a1ea-be6689e607ed","Type":"ContainerDied","Data":"40f7dc7f29d9aa9488c1ee47ca18e78ac062b2d82976a34fcdbda5c0ee3df075"} Oct 02 16:56:29 crc kubenswrapper[4808]: I1002 16:56:29.119177 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-twjnb"] Oct 02 16:56:29 crc kubenswrapper[4808]: I1002 16:56:29.240705 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-5fthx"] Oct 02 16:56:29 crc kubenswrapper[4808]: W1002 16:56:29.246246 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b6c4102_3ccd_4119_bbad_697ad2589e40.slice/crio-c41cdef9f12fedf5b245273a95b12a0147595137ea5b6d83558d3479132c405c WatchSource:0}: Error finding container c41cdef9f12fedf5b245273a95b12a0147595137ea5b6d83558d3479132c405c: Status 404 returned error can't find the container with id c41cdef9f12fedf5b245273a95b12a0147595137ea5b6d83558d3479132c405c Oct 02 16:56:29 crc kubenswrapper[4808]: I1002 16:56:29.956991 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f71818d6-1ebf-4838-b871-8873c107b325","Type":"ContainerStarted","Data":"c4f8b08ae8e573d59cb3204c0cdb27282c5471beafd2a27b8cc8c7464f5d768a"} Oct 02 16:56:29 crc kubenswrapper[4808]: I1002 16:56:29.957186 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 02 16:56:29 crc kubenswrapper[4808]: I1002 16:56:29.958219 4808 generic.go:334] "Generic (PLEG): container finished" podID="06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c" containerID="1a8aed86891bb916834343bdd5ae0633a486bc125740a2ec7873690cdf85f01a" exitCode=0 Oct 02 16:56:29 crc kubenswrapper[4808]: I1002 16:56:29.958286 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-twjnb" event={"ID":"06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c","Type":"ContainerDied","Data":"1a8aed86891bb916834343bdd5ae0633a486bc125740a2ec7873690cdf85f01a"} Oct 02 16:56:29 crc kubenswrapper[4808]: I1002 16:56:29.958310 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-twjnb" event={"ID":"06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c","Type":"ContainerStarted","Data":"b519ab8c1245429d06ce46125a2992ce7dd0a24cf90f7415cdfc13b1d1a5cb2c"} Oct 02 16:56:29 crc kubenswrapper[4808]: I1002 16:56:29.959509 4808 generic.go:334] "Generic (PLEG): container finished" podID="0b6c4102-3ccd-4119-bbad-697ad2589e40" containerID="60dbecb997aa51947969f39a54cbe1e672c3a01fe39c6255ff7e2d99445e9de4" exitCode=0 Oct 02 16:56:29 crc kubenswrapper[4808]: I1002 16:56:29.959554 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5fthx" event={"ID":"0b6c4102-3ccd-4119-bbad-697ad2589e40","Type":"ContainerDied","Data":"60dbecb997aa51947969f39a54cbe1e672c3a01fe39c6255ff7e2d99445e9de4"} Oct 02 16:56:29 crc kubenswrapper[4808]: I1002 16:56:29.959569 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5fthx" event={"ID":"0b6c4102-3ccd-4119-bbad-697ad2589e40","Type":"ContainerStarted","Data":"c41cdef9f12fedf5b245273a95b12a0147595137ea5b6d83558d3479132c405c"} Oct 02 16:56:29 crc kubenswrapper[4808]: I1002 16:56:29.963177 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b409b3d8-0ed6-49e9-a1ea-be6689e607ed","Type":"ContainerStarted","Data":"516b218bf929f18eb0453704db63a9a632bf10bea5cd9e6b24030634967a64fa"} Oct 02 16:56:29 crc kubenswrapper[4808]: I1002 16:56:29.963397 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:56:29 crc kubenswrapper[4808]: I1002 16:56:29.972054 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-hzbv4" podUID="eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8" containerName="ovn-controller" probeResult="failure" output=< Oct 02 16:56:29 crc kubenswrapper[4808]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 02 16:56:29 crc kubenswrapper[4808]: > Oct 02 16:56:29 crc kubenswrapper[4808]: I1002 16:56:29.977329 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.011932 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-59c8c" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.046907 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=49.078209711 podStartE2EDuration="57.046890475s" podCreationTimestamp="2025-10-02 16:55:33 +0000 UTC" firstStartedPulling="2025-10-02 16:55:47.205649609 +0000 UTC m=+914.531178609" lastFinishedPulling="2025-10-02 16:55:55.174330373 +0000 UTC m=+922.499859373" observedRunningTime="2025-10-02 16:56:30.0437674 +0000 UTC m=+957.369296400" watchObservedRunningTime="2025-10-02 16:56:30.046890475 +0000 UTC m=+957.372419475" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.061786 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=46.411340621 podStartE2EDuration="56.061766907s" podCreationTimestamp="2025-10-02 16:55:34 +0000 UTC" firstStartedPulling="2025-10-02 16:55:46.307314727 +0000 UTC m=+913.632843727" lastFinishedPulling="2025-10-02 16:55:55.957741003 +0000 UTC m=+923.283270013" observedRunningTime="2025-10-02 16:56:30.060486272 +0000 UTC m=+957.386015272" watchObservedRunningTime="2025-10-02 16:56:30.061766907 +0000 UTC m=+957.387295917" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.284105 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-hzbv4-config-gv2gn"] Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.286344 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.289670 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-hzbv4-config-gv2gn"] Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.290047 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.429915 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/10943164-fea8-4a30-a90e-78cf50d5eaa2-var-log-ovn\") pod \"ovn-controller-hzbv4-config-gv2gn\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.429987 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/10943164-fea8-4a30-a90e-78cf50d5eaa2-var-run\") pod \"ovn-controller-hzbv4-config-gv2gn\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.430013 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/10943164-fea8-4a30-a90e-78cf50d5eaa2-var-run-ovn\") pod \"ovn-controller-hzbv4-config-gv2gn\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.430074 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgdnl\" (UniqueName: \"kubernetes.io/projected/10943164-fea8-4a30-a90e-78cf50d5eaa2-kube-api-access-zgdnl\") pod \"ovn-controller-hzbv4-config-gv2gn\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.430105 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10943164-fea8-4a30-a90e-78cf50d5eaa2-scripts\") pod \"ovn-controller-hzbv4-config-gv2gn\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.430126 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/10943164-fea8-4a30-a90e-78cf50d5eaa2-additional-scripts\") pod \"ovn-controller-hzbv4-config-gv2gn\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.531167 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/10943164-fea8-4a30-a90e-78cf50d5eaa2-var-run\") pod \"ovn-controller-hzbv4-config-gv2gn\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.531225 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/10943164-fea8-4a30-a90e-78cf50d5eaa2-var-run-ovn\") pod \"ovn-controller-hzbv4-config-gv2gn\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.531294 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgdnl\" (UniqueName: \"kubernetes.io/projected/10943164-fea8-4a30-a90e-78cf50d5eaa2-kube-api-access-zgdnl\") pod \"ovn-controller-hzbv4-config-gv2gn\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.531323 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10943164-fea8-4a30-a90e-78cf50d5eaa2-scripts\") pod \"ovn-controller-hzbv4-config-gv2gn\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.531347 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/10943164-fea8-4a30-a90e-78cf50d5eaa2-additional-scripts\") pod \"ovn-controller-hzbv4-config-gv2gn\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.531454 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/10943164-fea8-4a30-a90e-78cf50d5eaa2-var-log-ovn\") pod \"ovn-controller-hzbv4-config-gv2gn\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.531507 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/10943164-fea8-4a30-a90e-78cf50d5eaa2-var-run\") pod \"ovn-controller-hzbv4-config-gv2gn\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.531549 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/10943164-fea8-4a30-a90e-78cf50d5eaa2-var-run-ovn\") pod \"ovn-controller-hzbv4-config-gv2gn\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.531578 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/10943164-fea8-4a30-a90e-78cf50d5eaa2-var-log-ovn\") pod \"ovn-controller-hzbv4-config-gv2gn\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.532305 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/10943164-fea8-4a30-a90e-78cf50d5eaa2-additional-scripts\") pod \"ovn-controller-hzbv4-config-gv2gn\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.533306 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10943164-fea8-4a30-a90e-78cf50d5eaa2-scripts\") pod \"ovn-controller-hzbv4-config-gv2gn\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.553610 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgdnl\" (UniqueName: \"kubernetes.io/projected/10943164-fea8-4a30-a90e-78cf50d5eaa2-kube-api-access-zgdnl\") pod \"ovn-controller-hzbv4-config-gv2gn\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.636971 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:30 crc kubenswrapper[4808]: I1002 16:56:30.987341 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-hzbv4-config-gv2gn"] Oct 02 16:56:31 crc kubenswrapper[4808]: I1002 16:56:31.254474 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5fthx" Oct 02 16:56:31 crc kubenswrapper[4808]: I1002 16:56:31.336896 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-twjnb" Oct 02 16:56:31 crc kubenswrapper[4808]: I1002 16:56:31.354867 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htxbm\" (UniqueName: \"kubernetes.io/projected/0b6c4102-3ccd-4119-bbad-697ad2589e40-kube-api-access-htxbm\") pod \"0b6c4102-3ccd-4119-bbad-697ad2589e40\" (UID: \"0b6c4102-3ccd-4119-bbad-697ad2589e40\") " Oct 02 16:56:31 crc kubenswrapper[4808]: I1002 16:56:31.366999 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b6c4102-3ccd-4119-bbad-697ad2589e40-kube-api-access-htxbm" (OuterVolumeSpecName: "kube-api-access-htxbm") pod "0b6c4102-3ccd-4119-bbad-697ad2589e40" (UID: "0b6c4102-3ccd-4119-bbad-697ad2589e40"). InnerVolumeSpecName "kube-api-access-htxbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:56:31 crc kubenswrapper[4808]: I1002 16:56:31.455927 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wpbc\" (UniqueName: \"kubernetes.io/projected/06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c-kube-api-access-6wpbc\") pod \"06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c\" (UID: \"06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c\") " Oct 02 16:56:31 crc kubenswrapper[4808]: I1002 16:56:31.456284 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htxbm\" (UniqueName: \"kubernetes.io/projected/0b6c4102-3ccd-4119-bbad-697ad2589e40-kube-api-access-htxbm\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:31 crc kubenswrapper[4808]: I1002 16:56:31.459671 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c-kube-api-access-6wpbc" (OuterVolumeSpecName: "kube-api-access-6wpbc") pod "06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c" (UID: "06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c"). InnerVolumeSpecName "kube-api-access-6wpbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:56:31 crc kubenswrapper[4808]: I1002 16:56:31.558414 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wpbc\" (UniqueName: \"kubernetes.io/projected/06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c-kube-api-access-6wpbc\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:31 crc kubenswrapper[4808]: I1002 16:56:31.981254 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-twjnb" event={"ID":"06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c","Type":"ContainerDied","Data":"b519ab8c1245429d06ce46125a2992ce7dd0a24cf90f7415cdfc13b1d1a5cb2c"} Oct 02 16:56:31 crc kubenswrapper[4808]: I1002 16:56:31.981309 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b519ab8c1245429d06ce46125a2992ce7dd0a24cf90f7415cdfc13b1d1a5cb2c" Oct 02 16:56:31 crc kubenswrapper[4808]: I1002 16:56:31.981328 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-twjnb" Oct 02 16:56:31 crc kubenswrapper[4808]: I1002 16:56:31.990070 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hzbv4-config-gv2gn" event={"ID":"10943164-fea8-4a30-a90e-78cf50d5eaa2","Type":"ContainerStarted","Data":"b46fa00211cf2548288c47a6e3f0f2d480b75fcb70fd1028ae56f9d300ce5ecc"} Oct 02 16:56:31 crc kubenswrapper[4808]: I1002 16:56:31.990115 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hzbv4-config-gv2gn" event={"ID":"10943164-fea8-4a30-a90e-78cf50d5eaa2","Type":"ContainerStarted","Data":"316549f8d88ddde12e77c495a3445cfb46aeef47a803fcddcf729f10b0a57925"} Oct 02 16:56:31 crc kubenswrapper[4808]: I1002 16:56:31.993147 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5fthx" event={"ID":"0b6c4102-3ccd-4119-bbad-697ad2589e40","Type":"ContainerDied","Data":"c41cdef9f12fedf5b245273a95b12a0147595137ea5b6d83558d3479132c405c"} Oct 02 16:56:31 crc kubenswrapper[4808]: I1002 16:56:31.993225 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c41cdef9f12fedf5b245273a95b12a0147595137ea5b6d83558d3479132c405c" Oct 02 16:56:31 crc kubenswrapper[4808]: I1002 16:56:31.993196 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5fthx" Oct 02 16:56:32 crc kubenswrapper[4808]: I1002 16:56:32.014687 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-hzbv4-config-gv2gn" podStartSLOduration=2.014666535 podStartE2EDuration="2.014666535s" podCreationTimestamp="2025-10-02 16:56:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:56:32.007064049 +0000 UTC m=+959.332593069" watchObservedRunningTime="2025-10-02 16:56:32.014666535 +0000 UTC m=+959.340195535" Oct 02 16:56:33 crc kubenswrapper[4808]: I1002 16:56:33.005361 4808 generic.go:334] "Generic (PLEG): container finished" podID="10943164-fea8-4a30-a90e-78cf50d5eaa2" containerID="b46fa00211cf2548288c47a6e3f0f2d480b75fcb70fd1028ae56f9d300ce5ecc" exitCode=0 Oct 02 16:56:33 crc kubenswrapper[4808]: I1002 16:56:33.005430 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hzbv4-config-gv2gn" event={"ID":"10943164-fea8-4a30-a90e-78cf50d5eaa2","Type":"ContainerDied","Data":"b46fa00211cf2548288c47a6e3f0f2d480b75fcb70fd1028ae56f9d300ce5ecc"} Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.000520 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-b4e3-account-create-qsp8f"] Oct 02 16:56:34 crc kubenswrapper[4808]: E1002 16:56:34.001147 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c" containerName="mariadb-database-create" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.001170 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c" containerName="mariadb-database-create" Oct 02 16:56:34 crc kubenswrapper[4808]: E1002 16:56:34.001189 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b6c4102-3ccd-4119-bbad-697ad2589e40" containerName="mariadb-database-create" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.001198 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b6c4102-3ccd-4119-bbad-697ad2589e40" containerName="mariadb-database-create" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.001367 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c" containerName="mariadb-database-create" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.001383 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b6c4102-3ccd-4119-bbad-697ad2589e40" containerName="mariadb-database-create" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.001991 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b4e3-account-create-qsp8f" Oct 02 16:56:34 crc kubenswrapper[4808]: W1002 16:56:34.003995 4808 reflector.go:561] object-"openstack"/"glance-db-secret": failed to list *v1.Secret: secrets "glance-db-secret" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 02 16:56:34 crc kubenswrapper[4808]: E1002 16:56:34.004057 4808 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"glance-db-secret\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"glance-db-secret\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.029836 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b4e3-account-create-qsp8f"] Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.103511 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jwsz\" (UniqueName: \"kubernetes.io/projected/387ae2f5-c39e-44a8-b471-64ab91a40d01-kube-api-access-7jwsz\") pod \"glance-b4e3-account-create-qsp8f\" (UID: \"387ae2f5-c39e-44a8-b471-64ab91a40d01\") " pod="openstack/glance-b4e3-account-create-qsp8f" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.206269 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jwsz\" (UniqueName: \"kubernetes.io/projected/387ae2f5-c39e-44a8-b471-64ab91a40d01-kube-api-access-7jwsz\") pod \"glance-b4e3-account-create-qsp8f\" (UID: \"387ae2f5-c39e-44a8-b471-64ab91a40d01\") " pod="openstack/glance-b4e3-account-create-qsp8f" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.248135 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jwsz\" (UniqueName: \"kubernetes.io/projected/387ae2f5-c39e-44a8-b471-64ab91a40d01-kube-api-access-7jwsz\") pod \"glance-b4e3-account-create-qsp8f\" (UID: \"387ae2f5-c39e-44a8-b471-64ab91a40d01\") " pod="openstack/glance-b4e3-account-create-qsp8f" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.364057 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b4e3-account-create-qsp8f" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.462952 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.612051 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10943164-fea8-4a30-a90e-78cf50d5eaa2-scripts\") pod \"10943164-fea8-4a30-a90e-78cf50d5eaa2\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.612101 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/10943164-fea8-4a30-a90e-78cf50d5eaa2-var-log-ovn\") pod \"10943164-fea8-4a30-a90e-78cf50d5eaa2\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.612154 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/10943164-fea8-4a30-a90e-78cf50d5eaa2-var-run-ovn\") pod \"10943164-fea8-4a30-a90e-78cf50d5eaa2\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.612191 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdnl\" (UniqueName: \"kubernetes.io/projected/10943164-fea8-4a30-a90e-78cf50d5eaa2-kube-api-access-zgdnl\") pod \"10943164-fea8-4a30-a90e-78cf50d5eaa2\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.612261 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/10943164-fea8-4a30-a90e-78cf50d5eaa2-additional-scripts\") pod \"10943164-fea8-4a30-a90e-78cf50d5eaa2\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.612284 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/10943164-fea8-4a30-a90e-78cf50d5eaa2-var-run\") pod \"10943164-fea8-4a30-a90e-78cf50d5eaa2\" (UID: \"10943164-fea8-4a30-a90e-78cf50d5eaa2\") " Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.612372 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10943164-fea8-4a30-a90e-78cf50d5eaa2-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "10943164-fea8-4a30-a90e-78cf50d5eaa2" (UID: "10943164-fea8-4a30-a90e-78cf50d5eaa2"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.612426 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10943164-fea8-4a30-a90e-78cf50d5eaa2-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "10943164-fea8-4a30-a90e-78cf50d5eaa2" (UID: "10943164-fea8-4a30-a90e-78cf50d5eaa2"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.612513 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10943164-fea8-4a30-a90e-78cf50d5eaa2-var-run" (OuterVolumeSpecName: "var-run") pod "10943164-fea8-4a30-a90e-78cf50d5eaa2" (UID: "10943164-fea8-4a30-a90e-78cf50d5eaa2"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.612684 4808 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/10943164-fea8-4a30-a90e-78cf50d5eaa2-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.612700 4808 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/10943164-fea8-4a30-a90e-78cf50d5eaa2-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.612711 4808 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/10943164-fea8-4a30-a90e-78cf50d5eaa2-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.613033 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10943164-fea8-4a30-a90e-78cf50d5eaa2-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "10943164-fea8-4a30-a90e-78cf50d5eaa2" (UID: "10943164-fea8-4a30-a90e-78cf50d5eaa2"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.614078 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10943164-fea8-4a30-a90e-78cf50d5eaa2-scripts" (OuterVolumeSpecName: "scripts") pod "10943164-fea8-4a30-a90e-78cf50d5eaa2" (UID: "10943164-fea8-4a30-a90e-78cf50d5eaa2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.628743 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10943164-fea8-4a30-a90e-78cf50d5eaa2-kube-api-access-zgdnl" (OuterVolumeSpecName: "kube-api-access-zgdnl") pod "10943164-fea8-4a30-a90e-78cf50d5eaa2" (UID: "10943164-fea8-4a30-a90e-78cf50d5eaa2"). InnerVolumeSpecName "kube-api-access-zgdnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.714619 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdnl\" (UniqueName: \"kubernetes.io/projected/10943164-fea8-4a30-a90e-78cf50d5eaa2-kube-api-access-zgdnl\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.714671 4808 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/10943164-fea8-4a30-a90e-78cf50d5eaa2-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.714692 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10943164-fea8-4a30-a90e-78cf50d5eaa2-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.790523 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b4e3-account-create-qsp8f"] Oct 02 16:56:34 crc kubenswrapper[4808]: I1002 16:56:34.900583 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-hzbv4" Oct 02 16:56:35 crc kubenswrapper[4808]: I1002 16:56:35.021039 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b4e3-account-create-qsp8f" event={"ID":"387ae2f5-c39e-44a8-b471-64ab91a40d01","Type":"ContainerStarted","Data":"5ad2aac829d076c40e54753c216a90f049d71c6efdb5ffba6e0209832e753e9d"} Oct 02 16:56:35 crc kubenswrapper[4808]: I1002 16:56:35.023626 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hzbv4-config-gv2gn" event={"ID":"10943164-fea8-4a30-a90e-78cf50d5eaa2","Type":"ContainerDied","Data":"316549f8d88ddde12e77c495a3445cfb46aeef47a803fcddcf729f10b0a57925"} Oct 02 16:56:35 crc kubenswrapper[4808]: I1002 16:56:35.023664 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="316549f8d88ddde12e77c495a3445cfb46aeef47a803fcddcf729f10b0a57925" Oct 02 16:56:35 crc kubenswrapper[4808]: I1002 16:56:35.023751 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hzbv4-config-gv2gn" Oct 02 16:56:35 crc kubenswrapper[4808]: I1002 16:56:35.174016 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-hzbv4-config-gv2gn"] Oct 02 16:56:35 crc kubenswrapper[4808]: I1002 16:56:35.183001 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-hzbv4-config-gv2gn"] Oct 02 16:56:35 crc kubenswrapper[4808]: I1002 16:56:35.282176 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 02 16:56:35 crc kubenswrapper[4808]: I1002 16:56:35.406219 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10943164-fea8-4a30-a90e-78cf50d5eaa2" path="/var/lib/kubelet/pods/10943164-fea8-4a30-a90e-78cf50d5eaa2/volumes" Oct 02 16:56:36 crc kubenswrapper[4808]: I1002 16:56:36.037517 4808 generic.go:334] "Generic (PLEG): container finished" podID="387ae2f5-c39e-44a8-b471-64ab91a40d01" containerID="aee639c7304a0f9c96eb2bba7121405627e0ac0f7f3c886fe1ee4d878b89acb4" exitCode=0 Oct 02 16:56:36 crc kubenswrapper[4808]: I1002 16:56:36.037582 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b4e3-account-create-qsp8f" event={"ID":"387ae2f5-c39e-44a8-b471-64ab91a40d01","Type":"ContainerDied","Data":"aee639c7304a0f9c96eb2bba7121405627e0ac0f7f3c886fe1ee4d878b89acb4"} Oct 02 16:56:37 crc kubenswrapper[4808]: I1002 16:56:37.407752 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b4e3-account-create-qsp8f" Oct 02 16:56:37 crc kubenswrapper[4808]: I1002 16:56:37.566124 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jwsz\" (UniqueName: \"kubernetes.io/projected/387ae2f5-c39e-44a8-b471-64ab91a40d01-kube-api-access-7jwsz\") pod \"387ae2f5-c39e-44a8-b471-64ab91a40d01\" (UID: \"387ae2f5-c39e-44a8-b471-64ab91a40d01\") " Oct 02 16:56:37 crc kubenswrapper[4808]: I1002 16:56:37.578632 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/387ae2f5-c39e-44a8-b471-64ab91a40d01-kube-api-access-7jwsz" (OuterVolumeSpecName: "kube-api-access-7jwsz") pod "387ae2f5-c39e-44a8-b471-64ab91a40d01" (UID: "387ae2f5-c39e-44a8-b471-64ab91a40d01"). InnerVolumeSpecName "kube-api-access-7jwsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:56:37 crc kubenswrapper[4808]: I1002 16:56:37.668512 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jwsz\" (UniqueName: \"kubernetes.io/projected/387ae2f5-c39e-44a8-b471-64ab91a40d01-kube-api-access-7jwsz\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.056537 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b4e3-account-create-qsp8f" event={"ID":"387ae2f5-c39e-44a8-b471-64ab91a40d01","Type":"ContainerDied","Data":"5ad2aac829d076c40e54753c216a90f049d71c6efdb5ffba6e0209832e753e9d"} Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.056578 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ad2aac829d076c40e54753c216a90f049d71c6efdb5ffba6e0209832e753e9d" Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.056619 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b4e3-account-create-qsp8f" Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.408105 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-3d12-account-create-wqs6d"] Oct 02 16:56:38 crc kubenswrapper[4808]: E1002 16:56:38.408435 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10943164-fea8-4a30-a90e-78cf50d5eaa2" containerName="ovn-config" Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.408447 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="10943164-fea8-4a30-a90e-78cf50d5eaa2" containerName="ovn-config" Oct 02 16:56:38 crc kubenswrapper[4808]: E1002 16:56:38.408461 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="387ae2f5-c39e-44a8-b471-64ab91a40d01" containerName="mariadb-account-create" Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.408470 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="387ae2f5-c39e-44a8-b471-64ab91a40d01" containerName="mariadb-account-create" Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.408666 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="387ae2f5-c39e-44a8-b471-64ab91a40d01" containerName="mariadb-account-create" Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.408684 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="10943164-fea8-4a30-a90e-78cf50d5eaa2" containerName="ovn-config" Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.409378 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3d12-account-create-wqs6d" Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.415378 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.422638 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-3d12-account-create-wqs6d"] Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.585820 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tqqq\" (UniqueName: \"kubernetes.io/projected/f446b584-b5e2-48fa-9f7d-3b6ccdc92944-kube-api-access-9tqqq\") pod \"keystone-3d12-account-create-wqs6d\" (UID: \"f446b584-b5e2-48fa-9f7d-3b6ccdc92944\") " pod="openstack/keystone-3d12-account-create-wqs6d" Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.687634 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tqqq\" (UniqueName: \"kubernetes.io/projected/f446b584-b5e2-48fa-9f7d-3b6ccdc92944-kube-api-access-9tqqq\") pod \"keystone-3d12-account-create-wqs6d\" (UID: \"f446b584-b5e2-48fa-9f7d-3b6ccdc92944\") " pod="openstack/keystone-3d12-account-create-wqs6d" Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.718364 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tqqq\" (UniqueName: \"kubernetes.io/projected/f446b584-b5e2-48fa-9f7d-3b6ccdc92944-kube-api-access-9tqqq\") pod \"keystone-3d12-account-create-wqs6d\" (UID: \"f446b584-b5e2-48fa-9f7d-3b6ccdc92944\") " pod="openstack/keystone-3d12-account-create-wqs6d" Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.726891 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3d12-account-create-wqs6d" Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.776012 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-828c-account-create-6vmr9"] Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.781155 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-828c-account-create-6vmr9" Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.782985 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.788904 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-828c-account-create-6vmr9"] Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.890073 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkrhv\" (UniqueName: \"kubernetes.io/projected/bf2f1836-9826-4d1c-89ea-eb979de32e62-kube-api-access-lkrhv\") pod \"placement-828c-account-create-6vmr9\" (UID: \"bf2f1836-9826-4d1c-89ea-eb979de32e62\") " pod="openstack/placement-828c-account-create-6vmr9" Oct 02 16:56:38 crc kubenswrapper[4808]: I1002 16:56:38.991457 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkrhv\" (UniqueName: \"kubernetes.io/projected/bf2f1836-9826-4d1c-89ea-eb979de32e62-kube-api-access-lkrhv\") pod \"placement-828c-account-create-6vmr9\" (UID: \"bf2f1836-9826-4d1c-89ea-eb979de32e62\") " pod="openstack/placement-828c-account-create-6vmr9" Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.011842 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkrhv\" (UniqueName: \"kubernetes.io/projected/bf2f1836-9826-4d1c-89ea-eb979de32e62-kube-api-access-lkrhv\") pod \"placement-828c-account-create-6vmr9\" (UID: \"bf2f1836-9826-4d1c-89ea-eb979de32e62\") " pod="openstack/placement-828c-account-create-6vmr9" Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.052103 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-67d46"] Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.058071 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-67d46"] Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.058166 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-67d46" Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.061106 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.062047 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-zx2bk" Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.149653 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-828c-account-create-6vmr9" Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.195647 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-combined-ca-bundle\") pod \"glance-db-sync-67d46\" (UID: \"8a5d0b28-6ba9-4421-98b6-1c6280ed7937\") " pod="openstack/glance-db-sync-67d46" Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.195735 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwd2g\" (UniqueName: \"kubernetes.io/projected/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-kube-api-access-pwd2g\") pod \"glance-db-sync-67d46\" (UID: \"8a5d0b28-6ba9-4421-98b6-1c6280ed7937\") " pod="openstack/glance-db-sync-67d46" Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.195797 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-db-sync-config-data\") pod \"glance-db-sync-67d46\" (UID: \"8a5d0b28-6ba9-4421-98b6-1c6280ed7937\") " pod="openstack/glance-db-sync-67d46" Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.195880 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-config-data\") pod \"glance-db-sync-67d46\" (UID: \"8a5d0b28-6ba9-4421-98b6-1c6280ed7937\") " pod="openstack/glance-db-sync-67d46" Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.235317 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-3d12-account-create-wqs6d"] Oct 02 16:56:39 crc kubenswrapper[4808]: W1002 16:56:39.245012 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf446b584_b5e2_48fa_9f7d_3b6ccdc92944.slice/crio-c3c0f02b6e24e7d0868d3ade8e31c6297b17e4c66d830c31c024e6e15dd8b3fd WatchSource:0}: Error finding container c3c0f02b6e24e7d0868d3ade8e31c6297b17e4c66d830c31c024e6e15dd8b3fd: Status 404 returned error can't find the container with id c3c0f02b6e24e7d0868d3ade8e31c6297b17e4c66d830c31c024e6e15dd8b3fd Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.296887 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-config-data\") pod \"glance-db-sync-67d46\" (UID: \"8a5d0b28-6ba9-4421-98b6-1c6280ed7937\") " pod="openstack/glance-db-sync-67d46" Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.297019 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-combined-ca-bundle\") pod \"glance-db-sync-67d46\" (UID: \"8a5d0b28-6ba9-4421-98b6-1c6280ed7937\") " pod="openstack/glance-db-sync-67d46" Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.297079 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwd2g\" (UniqueName: \"kubernetes.io/projected/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-kube-api-access-pwd2g\") pod \"glance-db-sync-67d46\" (UID: \"8a5d0b28-6ba9-4421-98b6-1c6280ed7937\") " pod="openstack/glance-db-sync-67d46" Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.297126 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-db-sync-config-data\") pod \"glance-db-sync-67d46\" (UID: \"8a5d0b28-6ba9-4421-98b6-1c6280ed7937\") " pod="openstack/glance-db-sync-67d46" Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.303958 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-config-data\") pod \"glance-db-sync-67d46\" (UID: \"8a5d0b28-6ba9-4421-98b6-1c6280ed7937\") " pod="openstack/glance-db-sync-67d46" Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.308346 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-db-sync-config-data\") pod \"glance-db-sync-67d46\" (UID: \"8a5d0b28-6ba9-4421-98b6-1c6280ed7937\") " pod="openstack/glance-db-sync-67d46" Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.311216 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-combined-ca-bundle\") pod \"glance-db-sync-67d46\" (UID: \"8a5d0b28-6ba9-4421-98b6-1c6280ed7937\") " pod="openstack/glance-db-sync-67d46" Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.315022 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwd2g\" (UniqueName: \"kubernetes.io/projected/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-kube-api-access-pwd2g\") pod \"glance-db-sync-67d46\" (UID: \"8a5d0b28-6ba9-4421-98b6-1c6280ed7937\") " pod="openstack/glance-db-sync-67d46" Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.385278 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-67d46" Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.422635 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-828c-account-create-6vmr9"] Oct 02 16:56:39 crc kubenswrapper[4808]: W1002 16:56:39.429154 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf2f1836_9826_4d1c_89ea_eb979de32e62.slice/crio-d774f832ee75d5d0133d83cf4c4127518ed88553d20e549c6dfa7f5594548ead WatchSource:0}: Error finding container d774f832ee75d5d0133d83cf4c4127518ed88553d20e549c6dfa7f5594548ead: Status 404 returned error can't find the container with id d774f832ee75d5d0133d83cf4c4127518ed88553d20e549c6dfa7f5594548ead Oct 02 16:56:39 crc kubenswrapper[4808]: I1002 16:56:39.991606 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-67d46"] Oct 02 16:56:40 crc kubenswrapper[4808]: I1002 16:56:40.076085 4808 generic.go:334] "Generic (PLEG): container finished" podID="bf2f1836-9826-4d1c-89ea-eb979de32e62" containerID="9806de53560962c1aac0dcf113979f67d8523dc5a68ce13ea383eb31316defae" exitCode=0 Oct 02 16:56:40 crc kubenswrapper[4808]: I1002 16:56:40.076157 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-828c-account-create-6vmr9" event={"ID":"bf2f1836-9826-4d1c-89ea-eb979de32e62","Type":"ContainerDied","Data":"9806de53560962c1aac0dcf113979f67d8523dc5a68ce13ea383eb31316defae"} Oct 02 16:56:40 crc kubenswrapper[4808]: I1002 16:56:40.076224 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-828c-account-create-6vmr9" event={"ID":"bf2f1836-9826-4d1c-89ea-eb979de32e62","Type":"ContainerStarted","Data":"d774f832ee75d5d0133d83cf4c4127518ed88553d20e549c6dfa7f5594548ead"} Oct 02 16:56:40 crc kubenswrapper[4808]: I1002 16:56:40.079041 4808 generic.go:334] "Generic (PLEG): container finished" podID="f446b584-b5e2-48fa-9f7d-3b6ccdc92944" containerID="0c6096da033553125ede6049f433d8b25e70a63f4c7dd2ce9f4774135f9c39ef" exitCode=0 Oct 02 16:56:40 crc kubenswrapper[4808]: I1002 16:56:40.079138 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3d12-account-create-wqs6d" event={"ID":"f446b584-b5e2-48fa-9f7d-3b6ccdc92944","Type":"ContainerDied","Data":"0c6096da033553125ede6049f433d8b25e70a63f4c7dd2ce9f4774135f9c39ef"} Oct 02 16:56:40 crc kubenswrapper[4808]: I1002 16:56:40.079180 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3d12-account-create-wqs6d" event={"ID":"f446b584-b5e2-48fa-9f7d-3b6ccdc92944","Type":"ContainerStarted","Data":"c3c0f02b6e24e7d0868d3ade8e31c6297b17e4c66d830c31c024e6e15dd8b3fd"} Oct 02 16:56:40 crc kubenswrapper[4808]: I1002 16:56:40.081731 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-67d46" event={"ID":"8a5d0b28-6ba9-4421-98b6-1c6280ed7937","Type":"ContainerStarted","Data":"8dc1e926bb3272714adcf9d4a7d5e32662fe8e7f6b1c24bf83e170ba4433a1be"} Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.300405 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56cd855b9c-czhcf"] Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.302579 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56cd855b9c-czhcf" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.319719 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56cd855b9c-czhcf"] Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.395491 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56cd855b9c-czhcf"] Oct 02 16:56:41 crc kubenswrapper[4808]: E1002 16:56:41.396388 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-mltps ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-56cd855b9c-czhcf" podUID="c5a82907-978b-4f8b-8a7c-667b68b8c4e1" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.435550 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-dns-svc\") pod \"dnsmasq-dns-56cd855b9c-czhcf\" (UID: \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\") " pod="openstack/dnsmasq-dns-56cd855b9c-czhcf" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.435613 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-ovsdbserver-sb\") pod \"dnsmasq-dns-56cd855b9c-czhcf\" (UID: \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\") " pod="openstack/dnsmasq-dns-56cd855b9c-czhcf" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.435641 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-ovsdbserver-nb\") pod \"dnsmasq-dns-56cd855b9c-czhcf\" (UID: \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\") " pod="openstack/dnsmasq-dns-56cd855b9c-czhcf" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.435687 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-config\") pod \"dnsmasq-dns-56cd855b9c-czhcf\" (UID: \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\") " pod="openstack/dnsmasq-dns-56cd855b9c-czhcf" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.435715 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mltps\" (UniqueName: \"kubernetes.io/projected/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-kube-api-access-mltps\") pod \"dnsmasq-dns-56cd855b9c-czhcf\" (UID: \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\") " pod="openstack/dnsmasq-dns-56cd855b9c-czhcf" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.537226 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-ovsdbserver-sb\") pod \"dnsmasq-dns-56cd855b9c-czhcf\" (UID: \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\") " pod="openstack/dnsmasq-dns-56cd855b9c-czhcf" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.537306 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-ovsdbserver-nb\") pod \"dnsmasq-dns-56cd855b9c-czhcf\" (UID: \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\") " pod="openstack/dnsmasq-dns-56cd855b9c-czhcf" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.537381 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-config\") pod \"dnsmasq-dns-56cd855b9c-czhcf\" (UID: \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\") " pod="openstack/dnsmasq-dns-56cd855b9c-czhcf" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.537407 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mltps\" (UniqueName: \"kubernetes.io/projected/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-kube-api-access-mltps\") pod \"dnsmasq-dns-56cd855b9c-czhcf\" (UID: \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\") " pod="openstack/dnsmasq-dns-56cd855b9c-czhcf" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.537484 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-dns-svc\") pod \"dnsmasq-dns-56cd855b9c-czhcf\" (UID: \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\") " pod="openstack/dnsmasq-dns-56cd855b9c-czhcf" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.539205 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-ovsdbserver-sb\") pod \"dnsmasq-dns-56cd855b9c-czhcf\" (UID: \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\") " pod="openstack/dnsmasq-dns-56cd855b9c-czhcf" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.539899 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-ovsdbserver-nb\") pod \"dnsmasq-dns-56cd855b9c-czhcf\" (UID: \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\") " pod="openstack/dnsmasq-dns-56cd855b9c-czhcf" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.543386 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-config\") pod \"dnsmasq-dns-56cd855b9c-czhcf\" (UID: \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\") " pod="openstack/dnsmasq-dns-56cd855b9c-czhcf" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.543918 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-dns-svc\") pod \"dnsmasq-dns-56cd855b9c-czhcf\" (UID: \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\") " pod="openstack/dnsmasq-dns-56cd855b9c-czhcf" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.564121 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3d12-account-create-wqs6d" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.587157 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mltps\" (UniqueName: \"kubernetes.io/projected/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-kube-api-access-mltps\") pod \"dnsmasq-dns-56cd855b9c-czhcf\" (UID: \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\") " pod="openstack/dnsmasq-dns-56cd855b9c-czhcf" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.663972 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tqqq\" (UniqueName: \"kubernetes.io/projected/f446b584-b5e2-48fa-9f7d-3b6ccdc92944-kube-api-access-9tqqq\") pod \"f446b584-b5e2-48fa-9f7d-3b6ccdc92944\" (UID: \"f446b584-b5e2-48fa-9f7d-3b6ccdc92944\") " Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.696855 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-828c-account-create-6vmr9" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.723445 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f446b584-b5e2-48fa-9f7d-3b6ccdc92944-kube-api-access-9tqqq" (OuterVolumeSpecName: "kube-api-access-9tqqq") pod "f446b584-b5e2-48fa-9f7d-3b6ccdc92944" (UID: "f446b584-b5e2-48fa-9f7d-3b6ccdc92944"). InnerVolumeSpecName "kube-api-access-9tqqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.806081 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkrhv\" (UniqueName: \"kubernetes.io/projected/bf2f1836-9826-4d1c-89ea-eb979de32e62-kube-api-access-lkrhv\") pod \"bf2f1836-9826-4d1c-89ea-eb979de32e62\" (UID: \"bf2f1836-9826-4d1c-89ea-eb979de32e62\") " Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.806458 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tqqq\" (UniqueName: \"kubernetes.io/projected/f446b584-b5e2-48fa-9f7d-3b6ccdc92944-kube-api-access-9tqqq\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.808953 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf2f1836-9826-4d1c-89ea-eb979de32e62-kube-api-access-lkrhv" (OuterVolumeSpecName: "kube-api-access-lkrhv") pod "bf2f1836-9826-4d1c-89ea-eb979de32e62" (UID: "bf2f1836-9826-4d1c-89ea-eb979de32e62"). InnerVolumeSpecName "kube-api-access-lkrhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:56:41 crc kubenswrapper[4808]: I1002 16:56:41.908098 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkrhv\" (UniqueName: \"kubernetes.io/projected/bf2f1836-9826-4d1c-89ea-eb979de32e62-kube-api-access-lkrhv\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.097943 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-828c-account-create-6vmr9" event={"ID":"bf2f1836-9826-4d1c-89ea-eb979de32e62","Type":"ContainerDied","Data":"d774f832ee75d5d0133d83cf4c4127518ed88553d20e549c6dfa7f5594548ead"} Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.097978 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d774f832ee75d5d0133d83cf4c4127518ed88553d20e549c6dfa7f5594548ead" Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.097985 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-828c-account-create-6vmr9" Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.099553 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3d12-account-create-wqs6d" Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.099592 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56cd855b9c-czhcf" Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.099578 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3d12-account-create-wqs6d" event={"ID":"f446b584-b5e2-48fa-9f7d-3b6ccdc92944","Type":"ContainerDied","Data":"c3c0f02b6e24e7d0868d3ade8e31c6297b17e4c66d830c31c024e6e15dd8b3fd"} Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.099627 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3c0f02b6e24e7d0868d3ade8e31c6297b17e4c66d830c31c024e6e15dd8b3fd" Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.109491 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56cd855b9c-czhcf" Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.211581 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-ovsdbserver-sb\") pod \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\" (UID: \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\") " Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.211621 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-config\") pod \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\" (UID: \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\") " Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.211714 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mltps\" (UniqueName: \"kubernetes.io/projected/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-kube-api-access-mltps\") pod \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\" (UID: \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\") " Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.211744 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-dns-svc\") pod \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\" (UID: \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\") " Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.211779 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-ovsdbserver-nb\") pod \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\" (UID: \"c5a82907-978b-4f8b-8a7c-667b68b8c4e1\") " Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.212639 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c5a82907-978b-4f8b-8a7c-667b68b8c4e1" (UID: "c5a82907-978b-4f8b-8a7c-667b68b8c4e1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.213221 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c5a82907-978b-4f8b-8a7c-667b68b8c4e1" (UID: "c5a82907-978b-4f8b-8a7c-667b68b8c4e1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.213773 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-config" (OuterVolumeSpecName: "config") pod "c5a82907-978b-4f8b-8a7c-667b68b8c4e1" (UID: "c5a82907-978b-4f8b-8a7c-667b68b8c4e1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.214023 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c5a82907-978b-4f8b-8a7c-667b68b8c4e1" (UID: "c5a82907-978b-4f8b-8a7c-667b68b8c4e1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.219861 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-kube-api-access-mltps" (OuterVolumeSpecName: "kube-api-access-mltps") pod "c5a82907-978b-4f8b-8a7c-667b68b8c4e1" (UID: "c5a82907-978b-4f8b-8a7c-667b68b8c4e1"). InnerVolumeSpecName "kube-api-access-mltps". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.313351 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.313381 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.313392 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mltps\" (UniqueName: \"kubernetes.io/projected/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-kube-api-access-mltps\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.313402 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:42 crc kubenswrapper[4808]: I1002 16:56:42.313411 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c5a82907-978b-4f8b-8a7c-667b68b8c4e1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:43 crc kubenswrapper[4808]: I1002 16:56:43.108331 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56cd855b9c-czhcf" Oct 02 16:56:43 crc kubenswrapper[4808]: I1002 16:56:43.178197 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56cd855b9c-czhcf"] Oct 02 16:56:43 crc kubenswrapper[4808]: I1002 16:56:43.192598 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56cd855b9c-czhcf"] Oct 02 16:56:43 crc kubenswrapper[4808]: I1002 16:56:43.424943 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5a82907-978b-4f8b-8a7c-667b68b8c4e1" path="/var/lib/kubelet/pods/c5a82907-978b-4f8b-8a7c-667b68b8c4e1/volumes" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.137202 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.409339 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-t6tjw"] Oct 02 16:56:45 crc kubenswrapper[4808]: E1002 16:56:45.409850 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf2f1836-9826-4d1c-89ea-eb979de32e62" containerName="mariadb-account-create" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.409866 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf2f1836-9826-4d1c-89ea-eb979de32e62" containerName="mariadb-account-create" Oct 02 16:56:45 crc kubenswrapper[4808]: E1002 16:56:45.409881 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f446b584-b5e2-48fa-9f7d-3b6ccdc92944" containerName="mariadb-account-create" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.409888 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f446b584-b5e2-48fa-9f7d-3b6ccdc92944" containerName="mariadb-account-create" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.410046 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf2f1836-9826-4d1c-89ea-eb979de32e62" containerName="mariadb-account-create" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.410059 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f446b584-b5e2-48fa-9f7d-3b6ccdc92944" containerName="mariadb-account-create" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.413037 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-t6tjw" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.433136 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-t6tjw"] Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.453427 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.469872 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pj89\" (UniqueName: \"kubernetes.io/projected/f2120107-7ced-4921-9a2d-eba436f36928-kube-api-access-8pj89\") pod \"barbican-db-create-t6tjw\" (UID: \"f2120107-7ced-4921-9a2d-eba436f36928\") " pod="openstack/barbican-db-create-t6tjw" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.573463 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pj89\" (UniqueName: \"kubernetes.io/projected/f2120107-7ced-4921-9a2d-eba436f36928-kube-api-access-8pj89\") pod \"barbican-db-create-t6tjw\" (UID: \"f2120107-7ced-4921-9a2d-eba436f36928\") " pod="openstack/barbican-db-create-t6tjw" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.592130 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-rr4z9"] Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.595406 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rr4z9" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.620537 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pj89\" (UniqueName: \"kubernetes.io/projected/f2120107-7ced-4921-9a2d-eba436f36928-kube-api-access-8pj89\") pod \"barbican-db-create-t6tjw\" (UID: \"f2120107-7ced-4921-9a2d-eba436f36928\") " pod="openstack/barbican-db-create-t6tjw" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.621255 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-rr4z9"] Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.677445 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zphvg\" (UniqueName: \"kubernetes.io/projected/af84bfb8-1583-46a8-ae8f-1b28b3568311-kube-api-access-zphvg\") pod \"cinder-db-create-rr4z9\" (UID: \"af84bfb8-1583-46a8-ae8f-1b28b3568311\") " pod="openstack/cinder-db-create-rr4z9" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.694792 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-6bh5l"] Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.695810 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6bh5l" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.705781 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-6bh5l"] Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.761214 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-7wz2m"] Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.762331 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-7wz2m" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.764502 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-b9ff9" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.764680 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.764782 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.764788 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.766174 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-7wz2m"] Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.776150 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-t6tjw" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.779395 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zphvg\" (UniqueName: \"kubernetes.io/projected/af84bfb8-1583-46a8-ae8f-1b28b3568311-kube-api-access-zphvg\") pod \"cinder-db-create-rr4z9\" (UID: \"af84bfb8-1583-46a8-ae8f-1b28b3568311\") " pod="openstack/cinder-db-create-rr4z9" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.779436 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk76c\" (UniqueName: \"kubernetes.io/projected/9f95dff9-cca1-4356-912a-76d1475a273a-kube-api-access-jk76c\") pod \"neutron-db-create-6bh5l\" (UID: \"9f95dff9-cca1-4356-912a-76d1475a273a\") " pod="openstack/neutron-db-create-6bh5l" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.797619 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zphvg\" (UniqueName: \"kubernetes.io/projected/af84bfb8-1583-46a8-ae8f-1b28b3568311-kube-api-access-zphvg\") pod \"cinder-db-create-rr4z9\" (UID: \"af84bfb8-1583-46a8-ae8f-1b28b3568311\") " pod="openstack/cinder-db-create-rr4z9" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.881353 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk76c\" (UniqueName: \"kubernetes.io/projected/9f95dff9-cca1-4356-912a-76d1475a273a-kube-api-access-jk76c\") pod \"neutron-db-create-6bh5l\" (UID: \"9f95dff9-cca1-4356-912a-76d1475a273a\") " pod="openstack/neutron-db-create-6bh5l" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.881888 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfeb1719-f48e-488b-8c22-a9be946819e3-combined-ca-bundle\") pod \"keystone-db-sync-7wz2m\" (UID: \"cfeb1719-f48e-488b-8c22-a9be946819e3\") " pod="openstack/keystone-db-sync-7wz2m" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.882008 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfeb1719-f48e-488b-8c22-a9be946819e3-config-data\") pod \"keystone-db-sync-7wz2m\" (UID: \"cfeb1719-f48e-488b-8c22-a9be946819e3\") " pod="openstack/keystone-db-sync-7wz2m" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.882090 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w24rs\" (UniqueName: \"kubernetes.io/projected/cfeb1719-f48e-488b-8c22-a9be946819e3-kube-api-access-w24rs\") pod \"keystone-db-sync-7wz2m\" (UID: \"cfeb1719-f48e-488b-8c22-a9be946819e3\") " pod="openstack/keystone-db-sync-7wz2m" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.895730 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk76c\" (UniqueName: \"kubernetes.io/projected/9f95dff9-cca1-4356-912a-76d1475a273a-kube-api-access-jk76c\") pod \"neutron-db-create-6bh5l\" (UID: \"9f95dff9-cca1-4356-912a-76d1475a273a\") " pod="openstack/neutron-db-create-6bh5l" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.955772 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rr4z9" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.983063 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfeb1719-f48e-488b-8c22-a9be946819e3-config-data\") pod \"keystone-db-sync-7wz2m\" (UID: \"cfeb1719-f48e-488b-8c22-a9be946819e3\") " pod="openstack/keystone-db-sync-7wz2m" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.983140 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w24rs\" (UniqueName: \"kubernetes.io/projected/cfeb1719-f48e-488b-8c22-a9be946819e3-kube-api-access-w24rs\") pod \"keystone-db-sync-7wz2m\" (UID: \"cfeb1719-f48e-488b-8c22-a9be946819e3\") " pod="openstack/keystone-db-sync-7wz2m" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.983203 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfeb1719-f48e-488b-8c22-a9be946819e3-combined-ca-bundle\") pod \"keystone-db-sync-7wz2m\" (UID: \"cfeb1719-f48e-488b-8c22-a9be946819e3\") " pod="openstack/keystone-db-sync-7wz2m" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.986413 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfeb1719-f48e-488b-8c22-a9be946819e3-config-data\") pod \"keystone-db-sync-7wz2m\" (UID: \"cfeb1719-f48e-488b-8c22-a9be946819e3\") " pod="openstack/keystone-db-sync-7wz2m" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.986434 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfeb1719-f48e-488b-8c22-a9be946819e3-combined-ca-bundle\") pod \"keystone-db-sync-7wz2m\" (UID: \"cfeb1719-f48e-488b-8c22-a9be946819e3\") " pod="openstack/keystone-db-sync-7wz2m" Oct 02 16:56:45 crc kubenswrapper[4808]: I1002 16:56:45.999876 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w24rs\" (UniqueName: \"kubernetes.io/projected/cfeb1719-f48e-488b-8c22-a9be946819e3-kube-api-access-w24rs\") pod \"keystone-db-sync-7wz2m\" (UID: \"cfeb1719-f48e-488b-8c22-a9be946819e3\") " pod="openstack/keystone-db-sync-7wz2m" Oct 02 16:56:46 crc kubenswrapper[4808]: I1002 16:56:46.011190 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6bh5l" Oct 02 16:56:46 crc kubenswrapper[4808]: I1002 16:56:46.085991 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-7wz2m" Oct 02 16:56:53 crc kubenswrapper[4808]: I1002 16:56:53.042007 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-t6tjw"] Oct 02 16:56:53 crc kubenswrapper[4808]: W1002 16:56:53.061091 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2120107_7ced_4921_9a2d_eba436f36928.slice/crio-3218cbb4c52ea48cdcf93a1d17dfbd69791d5480d98db7277032831f7fa1ca26 WatchSource:0}: Error finding container 3218cbb4c52ea48cdcf93a1d17dfbd69791d5480d98db7277032831f7fa1ca26: Status 404 returned error can't find the container with id 3218cbb4c52ea48cdcf93a1d17dfbd69791d5480d98db7277032831f7fa1ca26 Oct 02 16:56:53 crc kubenswrapper[4808]: I1002 16:56:53.142639 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-7wz2m"] Oct 02 16:56:53 crc kubenswrapper[4808]: I1002 16:56:53.147708 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-rr4z9"] Oct 02 16:56:53 crc kubenswrapper[4808]: W1002 16:56:53.160262 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfeb1719_f48e_488b_8c22_a9be946819e3.slice/crio-dfebc06ff291d3723b2b4d0f74c40fd34e07254748af9f3059ce479f5421f2e4 WatchSource:0}: Error finding container dfebc06ff291d3723b2b4d0f74c40fd34e07254748af9f3059ce479f5421f2e4: Status 404 returned error can't find the container with id dfebc06ff291d3723b2b4d0f74c40fd34e07254748af9f3059ce479f5421f2e4 Oct 02 16:56:53 crc kubenswrapper[4808]: W1002 16:56:53.164058 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf84bfb8_1583_46a8_ae8f_1b28b3568311.slice/crio-49a95724dc6eb5fdd79303f3fe99544f81d6b82629cb2eeb4ba016a79951f0a1 WatchSource:0}: Error finding container 49a95724dc6eb5fdd79303f3fe99544f81d6b82629cb2eeb4ba016a79951f0a1: Status 404 returned error can't find the container with id 49a95724dc6eb5fdd79303f3fe99544f81d6b82629cb2eeb4ba016a79951f0a1 Oct 02 16:56:53 crc kubenswrapper[4808]: I1002 16:56:53.221786 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-7wz2m" event={"ID":"cfeb1719-f48e-488b-8c22-a9be946819e3","Type":"ContainerStarted","Data":"dfebc06ff291d3723b2b4d0f74c40fd34e07254748af9f3059ce479f5421f2e4"} Oct 02 16:56:53 crc kubenswrapper[4808]: I1002 16:56:53.227082 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rr4z9" event={"ID":"af84bfb8-1583-46a8-ae8f-1b28b3568311","Type":"ContainerStarted","Data":"49a95724dc6eb5fdd79303f3fe99544f81d6b82629cb2eeb4ba016a79951f0a1"} Oct 02 16:56:53 crc kubenswrapper[4808]: I1002 16:56:53.228374 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-6bh5l"] Oct 02 16:56:53 crc kubenswrapper[4808]: I1002 16:56:53.232551 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-t6tjw" event={"ID":"f2120107-7ced-4921-9a2d-eba436f36928","Type":"ContainerStarted","Data":"3218cbb4c52ea48cdcf93a1d17dfbd69791d5480d98db7277032831f7fa1ca26"} Oct 02 16:56:53 crc kubenswrapper[4808]: W1002 16:56:53.239266 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f95dff9_cca1_4356_912a_76d1475a273a.slice/crio-789690b8c51e9d2050870416cdbc46f5404e0d43fabed94735cff4288e72f09b WatchSource:0}: Error finding container 789690b8c51e9d2050870416cdbc46f5404e0d43fabed94735cff4288e72f09b: Status 404 returned error can't find the container with id 789690b8c51e9d2050870416cdbc46f5404e0d43fabed94735cff4288e72f09b Oct 02 16:56:54 crc kubenswrapper[4808]: I1002 16:56:54.241882 4808 generic.go:334] "Generic (PLEG): container finished" podID="af84bfb8-1583-46a8-ae8f-1b28b3568311" containerID="82fe42107df11b8d00d292a2debac65bb0fcd09f7ebc87e92644a02e72e11349" exitCode=0 Oct 02 16:56:54 crc kubenswrapper[4808]: I1002 16:56:54.241989 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rr4z9" event={"ID":"af84bfb8-1583-46a8-ae8f-1b28b3568311","Type":"ContainerDied","Data":"82fe42107df11b8d00d292a2debac65bb0fcd09f7ebc87e92644a02e72e11349"} Oct 02 16:56:54 crc kubenswrapper[4808]: I1002 16:56:54.243661 4808 generic.go:334] "Generic (PLEG): container finished" podID="9f95dff9-cca1-4356-912a-76d1475a273a" containerID="5e43cb4708cd4c2e4969ef25f0191c4b0d425d70cbc4484986bd60e8b5a6bdc8" exitCode=0 Oct 02 16:56:54 crc kubenswrapper[4808]: I1002 16:56:54.243695 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6bh5l" event={"ID":"9f95dff9-cca1-4356-912a-76d1475a273a","Type":"ContainerDied","Data":"5e43cb4708cd4c2e4969ef25f0191c4b0d425d70cbc4484986bd60e8b5a6bdc8"} Oct 02 16:56:54 crc kubenswrapper[4808]: I1002 16:56:54.243763 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6bh5l" event={"ID":"9f95dff9-cca1-4356-912a-76d1475a273a","Type":"ContainerStarted","Data":"789690b8c51e9d2050870416cdbc46f5404e0d43fabed94735cff4288e72f09b"} Oct 02 16:56:54 crc kubenswrapper[4808]: I1002 16:56:54.247977 4808 generic.go:334] "Generic (PLEG): container finished" podID="f2120107-7ced-4921-9a2d-eba436f36928" containerID="f1b6080bd93c4a4a6d797f30ad2dff9e3464e35423aa67e539aba687184a0847" exitCode=0 Oct 02 16:56:54 crc kubenswrapper[4808]: I1002 16:56:54.248026 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-t6tjw" event={"ID":"f2120107-7ced-4921-9a2d-eba436f36928","Type":"ContainerDied","Data":"f1b6080bd93c4a4a6d797f30ad2dff9e3464e35423aa67e539aba687184a0847"} Oct 02 16:56:54 crc kubenswrapper[4808]: I1002 16:56:54.249731 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-67d46" event={"ID":"8a5d0b28-6ba9-4421-98b6-1c6280ed7937","Type":"ContainerStarted","Data":"8183cfd2823a8ad77830639cf768f14779529261544b212659991b9b6fbd628a"} Oct 02 16:56:54 crc kubenswrapper[4808]: I1002 16:56:54.282961 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-67d46" podStartSLOduration=2.660911098 podStartE2EDuration="15.282944262s" podCreationTimestamp="2025-10-02 16:56:39 +0000 UTC" firstStartedPulling="2025-10-02 16:56:40.011352096 +0000 UTC m=+967.336881106" lastFinishedPulling="2025-10-02 16:56:52.63338526 +0000 UTC m=+979.958914270" observedRunningTime="2025-10-02 16:56:54.279153389 +0000 UTC m=+981.604682439" watchObservedRunningTime="2025-10-02 16:56:54.282944262 +0000 UTC m=+981.608473272" Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.055043 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6bh5l" Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.097439 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rr4z9" Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.123972 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-t6tjw" Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.192655 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jk76c\" (UniqueName: \"kubernetes.io/projected/9f95dff9-cca1-4356-912a-76d1475a273a-kube-api-access-jk76c\") pod \"9f95dff9-cca1-4356-912a-76d1475a273a\" (UID: \"9f95dff9-cca1-4356-912a-76d1475a273a\") " Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.192771 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zphvg\" (UniqueName: \"kubernetes.io/projected/af84bfb8-1583-46a8-ae8f-1b28b3568311-kube-api-access-zphvg\") pod \"af84bfb8-1583-46a8-ae8f-1b28b3568311\" (UID: \"af84bfb8-1583-46a8-ae8f-1b28b3568311\") " Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.197608 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af84bfb8-1583-46a8-ae8f-1b28b3568311-kube-api-access-zphvg" (OuterVolumeSpecName: "kube-api-access-zphvg") pod "af84bfb8-1583-46a8-ae8f-1b28b3568311" (UID: "af84bfb8-1583-46a8-ae8f-1b28b3568311"). InnerVolumeSpecName "kube-api-access-zphvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.197998 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f95dff9-cca1-4356-912a-76d1475a273a-kube-api-access-jk76c" (OuterVolumeSpecName: "kube-api-access-jk76c") pod "9f95dff9-cca1-4356-912a-76d1475a273a" (UID: "9f95dff9-cca1-4356-912a-76d1475a273a"). InnerVolumeSpecName "kube-api-access-jk76c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.280639 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6bh5l" event={"ID":"9f95dff9-cca1-4356-912a-76d1475a273a","Type":"ContainerDied","Data":"789690b8c51e9d2050870416cdbc46f5404e0d43fabed94735cff4288e72f09b"} Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.280675 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="789690b8c51e9d2050870416cdbc46f5404e0d43fabed94735cff4288e72f09b" Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.280704 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6bh5l" Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.282529 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-t6tjw" Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.282534 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-t6tjw" event={"ID":"f2120107-7ced-4921-9a2d-eba436f36928","Type":"ContainerDied","Data":"3218cbb4c52ea48cdcf93a1d17dfbd69791d5480d98db7277032831f7fa1ca26"} Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.282591 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3218cbb4c52ea48cdcf93a1d17dfbd69791d5480d98db7277032831f7fa1ca26" Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.283816 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-7wz2m" event={"ID":"cfeb1719-f48e-488b-8c22-a9be946819e3","Type":"ContainerStarted","Data":"ca4cc9c311e0b366cbc723acdc339555b046721505c67fe0ca3c881888b76936"} Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.285724 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rr4z9" event={"ID":"af84bfb8-1583-46a8-ae8f-1b28b3568311","Type":"ContainerDied","Data":"49a95724dc6eb5fdd79303f3fe99544f81d6b82629cb2eeb4ba016a79951f0a1"} Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.285743 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49a95724dc6eb5fdd79303f3fe99544f81d6b82629cb2eeb4ba016a79951f0a1" Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.285765 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rr4z9" Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.295669 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pj89\" (UniqueName: \"kubernetes.io/projected/f2120107-7ced-4921-9a2d-eba436f36928-kube-api-access-8pj89\") pod \"f2120107-7ced-4921-9a2d-eba436f36928\" (UID: \"f2120107-7ced-4921-9a2d-eba436f36928\") " Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.296430 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jk76c\" (UniqueName: \"kubernetes.io/projected/9f95dff9-cca1-4356-912a-76d1475a273a-kube-api-access-jk76c\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.296471 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zphvg\" (UniqueName: \"kubernetes.io/projected/af84bfb8-1583-46a8-ae8f-1b28b3568311-kube-api-access-zphvg\") on node \"crc\" DevicePath \"\"" Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.299524 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2120107-7ced-4921-9a2d-eba436f36928-kube-api-access-8pj89" (OuterVolumeSpecName: "kube-api-access-8pj89") pod "f2120107-7ced-4921-9a2d-eba436f36928" (UID: "f2120107-7ced-4921-9a2d-eba436f36928"). InnerVolumeSpecName "kube-api-access-8pj89". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.312971 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-7wz2m" podStartSLOduration=8.515284819 podStartE2EDuration="13.3129511s" podCreationTimestamp="2025-10-02 16:56:45 +0000 UTC" firstStartedPulling="2025-10-02 16:56:53.163617443 +0000 UTC m=+980.489146443" lastFinishedPulling="2025-10-02 16:56:57.961283704 +0000 UTC m=+985.286812724" observedRunningTime="2025-10-02 16:56:58.302963729 +0000 UTC m=+985.628492749" watchObservedRunningTime="2025-10-02 16:56:58.3129511 +0000 UTC m=+985.638480100" Oct 02 16:56:58 crc kubenswrapper[4808]: I1002 16:56:58.397961 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pj89\" (UniqueName: \"kubernetes.io/projected/f2120107-7ced-4921-9a2d-eba436f36928-kube-api-access-8pj89\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:00 crc kubenswrapper[4808]: I1002 16:57:00.320317 4808 generic.go:334] "Generic (PLEG): container finished" podID="8a5d0b28-6ba9-4421-98b6-1c6280ed7937" containerID="8183cfd2823a8ad77830639cf768f14779529261544b212659991b9b6fbd628a" exitCode=0 Oct 02 16:57:00 crc kubenswrapper[4808]: I1002 16:57:00.320771 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-67d46" event={"ID":"8a5d0b28-6ba9-4421-98b6-1c6280ed7937","Type":"ContainerDied","Data":"8183cfd2823a8ad77830639cf768f14779529261544b212659991b9b6fbd628a"} Oct 02 16:57:01 crc kubenswrapper[4808]: I1002 16:57:01.337160 4808 generic.go:334] "Generic (PLEG): container finished" podID="cfeb1719-f48e-488b-8c22-a9be946819e3" containerID="ca4cc9c311e0b366cbc723acdc339555b046721505c67fe0ca3c881888b76936" exitCode=0 Oct 02 16:57:01 crc kubenswrapper[4808]: I1002 16:57:01.337647 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-7wz2m" event={"ID":"cfeb1719-f48e-488b-8c22-a9be946819e3","Type":"ContainerDied","Data":"ca4cc9c311e0b366cbc723acdc339555b046721505c67fe0ca3c881888b76936"} Oct 02 16:57:01 crc kubenswrapper[4808]: I1002 16:57:01.752500 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-67d46" Oct 02 16:57:01 crc kubenswrapper[4808]: I1002 16:57:01.857284 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwd2g\" (UniqueName: \"kubernetes.io/projected/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-kube-api-access-pwd2g\") pod \"8a5d0b28-6ba9-4421-98b6-1c6280ed7937\" (UID: \"8a5d0b28-6ba9-4421-98b6-1c6280ed7937\") " Oct 02 16:57:01 crc kubenswrapper[4808]: I1002 16:57:01.857449 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-config-data\") pod \"8a5d0b28-6ba9-4421-98b6-1c6280ed7937\" (UID: \"8a5d0b28-6ba9-4421-98b6-1c6280ed7937\") " Oct 02 16:57:01 crc kubenswrapper[4808]: I1002 16:57:01.857550 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-combined-ca-bundle\") pod \"8a5d0b28-6ba9-4421-98b6-1c6280ed7937\" (UID: \"8a5d0b28-6ba9-4421-98b6-1c6280ed7937\") " Oct 02 16:57:01 crc kubenswrapper[4808]: I1002 16:57:01.857584 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-db-sync-config-data\") pod \"8a5d0b28-6ba9-4421-98b6-1c6280ed7937\" (UID: \"8a5d0b28-6ba9-4421-98b6-1c6280ed7937\") " Oct 02 16:57:01 crc kubenswrapper[4808]: I1002 16:57:01.863895 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-kube-api-access-pwd2g" (OuterVolumeSpecName: "kube-api-access-pwd2g") pod "8a5d0b28-6ba9-4421-98b6-1c6280ed7937" (UID: "8a5d0b28-6ba9-4421-98b6-1c6280ed7937"). InnerVolumeSpecName "kube-api-access-pwd2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:57:01 crc kubenswrapper[4808]: I1002 16:57:01.867552 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8a5d0b28-6ba9-4421-98b6-1c6280ed7937" (UID: "8a5d0b28-6ba9-4421-98b6-1c6280ed7937"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:57:01 crc kubenswrapper[4808]: I1002 16:57:01.904052 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a5d0b28-6ba9-4421-98b6-1c6280ed7937" (UID: "8a5d0b28-6ba9-4421-98b6-1c6280ed7937"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:57:01 crc kubenswrapper[4808]: I1002 16:57:01.928641 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-config-data" (OuterVolumeSpecName: "config-data") pod "8a5d0b28-6ba9-4421-98b6-1c6280ed7937" (UID: "8a5d0b28-6ba9-4421-98b6-1c6280ed7937"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:57:01 crc kubenswrapper[4808]: I1002 16:57:01.959651 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:01 crc kubenswrapper[4808]: I1002 16:57:01.959690 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:01 crc kubenswrapper[4808]: I1002 16:57:01.959705 4808 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:01 crc kubenswrapper[4808]: I1002 16:57:01.959718 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwd2g\" (UniqueName: \"kubernetes.io/projected/8a5d0b28-6ba9-4421-98b6-1c6280ed7937-kube-api-access-pwd2g\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.347665 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-67d46" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.347663 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-67d46" event={"ID":"8a5d0b28-6ba9-4421-98b6-1c6280ed7937","Type":"ContainerDied","Data":"8dc1e926bb3272714adcf9d4a7d5e32662fe8e7f6b1c24bf83e170ba4433a1be"} Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.347727 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8dc1e926bb3272714adcf9d4a7d5e32662fe8e7f6b1c24bf83e170ba4433a1be" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.787459 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-7wz2m" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.805990 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-fgskc"] Oct 02 16:57:02 crc kubenswrapper[4808]: E1002 16:57:02.806501 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af84bfb8-1583-46a8-ae8f-1b28b3568311" containerName="mariadb-database-create" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.806527 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="af84bfb8-1583-46a8-ae8f-1b28b3568311" containerName="mariadb-database-create" Oct 02 16:57:02 crc kubenswrapper[4808]: E1002 16:57:02.806545 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f95dff9-cca1-4356-912a-76d1475a273a" containerName="mariadb-database-create" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.806558 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f95dff9-cca1-4356-912a-76d1475a273a" containerName="mariadb-database-create" Oct 02 16:57:02 crc kubenswrapper[4808]: E1002 16:57:02.806591 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfeb1719-f48e-488b-8c22-a9be946819e3" containerName="keystone-db-sync" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.806608 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfeb1719-f48e-488b-8c22-a9be946819e3" containerName="keystone-db-sync" Oct 02 16:57:02 crc kubenswrapper[4808]: E1002 16:57:02.806616 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a5d0b28-6ba9-4421-98b6-1c6280ed7937" containerName="glance-db-sync" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.806630 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a5d0b28-6ba9-4421-98b6-1c6280ed7937" containerName="glance-db-sync" Oct 02 16:57:02 crc kubenswrapper[4808]: E1002 16:57:02.806653 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2120107-7ced-4921-9a2d-eba436f36928" containerName="mariadb-database-create" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.806662 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2120107-7ced-4921-9a2d-eba436f36928" containerName="mariadb-database-create" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.806884 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="af84bfb8-1583-46a8-ae8f-1b28b3568311" containerName="mariadb-database-create" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.806907 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f95dff9-cca1-4356-912a-76d1475a273a" containerName="mariadb-database-create" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.806922 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2120107-7ced-4921-9a2d-eba436f36928" containerName="mariadb-database-create" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.806935 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfeb1719-f48e-488b-8c22-a9be946819e3" containerName="keystone-db-sync" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.806951 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a5d0b28-6ba9-4421-98b6-1c6280ed7937" containerName="glance-db-sync" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.807995 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.830605 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-fgskc"] Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.876920 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfeb1719-f48e-488b-8c22-a9be946819e3-config-data\") pod \"cfeb1719-f48e-488b-8c22-a9be946819e3\" (UID: \"cfeb1719-f48e-488b-8c22-a9be946819e3\") " Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.877081 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w24rs\" (UniqueName: \"kubernetes.io/projected/cfeb1719-f48e-488b-8c22-a9be946819e3-kube-api-access-w24rs\") pod \"cfeb1719-f48e-488b-8c22-a9be946819e3\" (UID: \"cfeb1719-f48e-488b-8c22-a9be946819e3\") " Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.877299 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfeb1719-f48e-488b-8c22-a9be946819e3-combined-ca-bundle\") pod \"cfeb1719-f48e-488b-8c22-a9be946819e3\" (UID: \"cfeb1719-f48e-488b-8c22-a9be946819e3\") " Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.896567 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfeb1719-f48e-488b-8c22-a9be946819e3-kube-api-access-w24rs" (OuterVolumeSpecName: "kube-api-access-w24rs") pod "cfeb1719-f48e-488b-8c22-a9be946819e3" (UID: "cfeb1719-f48e-488b-8c22-a9be946819e3"). InnerVolumeSpecName "kube-api-access-w24rs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.920733 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfeb1719-f48e-488b-8c22-a9be946819e3-config-data" (OuterVolumeSpecName: "config-data") pod "cfeb1719-f48e-488b-8c22-a9be946819e3" (UID: "cfeb1719-f48e-488b-8c22-a9be946819e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.946491 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfeb1719-f48e-488b-8c22-a9be946819e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cfeb1719-f48e-488b-8c22-a9be946819e3" (UID: "cfeb1719-f48e-488b-8c22-a9be946819e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.978648 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-ovsdbserver-sb\") pod \"dnsmasq-dns-54f9b7b8d9-fgskc\" (UID: \"9aca5301-55c6-4e78-8830-11cb79861ade\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.978723 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rbj4\" (UniqueName: \"kubernetes.io/projected/9aca5301-55c6-4e78-8830-11cb79861ade-kube-api-access-4rbj4\") pod \"dnsmasq-dns-54f9b7b8d9-fgskc\" (UID: \"9aca5301-55c6-4e78-8830-11cb79861ade\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.978772 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-dns-svc\") pod \"dnsmasq-dns-54f9b7b8d9-fgskc\" (UID: \"9aca5301-55c6-4e78-8830-11cb79861ade\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.978921 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-ovsdbserver-nb\") pod \"dnsmasq-dns-54f9b7b8d9-fgskc\" (UID: \"9aca5301-55c6-4e78-8830-11cb79861ade\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.978967 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-config\") pod \"dnsmasq-dns-54f9b7b8d9-fgskc\" (UID: \"9aca5301-55c6-4e78-8830-11cb79861ade\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.979187 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfeb1719-f48e-488b-8c22-a9be946819e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.979205 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfeb1719-f48e-488b-8c22-a9be946819e3-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:02 crc kubenswrapper[4808]: I1002 16:57:02.979216 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w24rs\" (UniqueName: \"kubernetes.io/projected/cfeb1719-f48e-488b-8c22-a9be946819e3-kube-api-access-w24rs\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:03 crc kubenswrapper[4808]: I1002 16:57:03.081026 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-dns-svc\") pod \"dnsmasq-dns-54f9b7b8d9-fgskc\" (UID: \"9aca5301-55c6-4e78-8830-11cb79861ade\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:57:03 crc kubenswrapper[4808]: I1002 16:57:03.081202 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-ovsdbserver-nb\") pod \"dnsmasq-dns-54f9b7b8d9-fgskc\" (UID: \"9aca5301-55c6-4e78-8830-11cb79861ade\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:57:03 crc kubenswrapper[4808]: I1002 16:57:03.081294 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-config\") pod \"dnsmasq-dns-54f9b7b8d9-fgskc\" (UID: \"9aca5301-55c6-4e78-8830-11cb79861ade\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:57:03 crc kubenswrapper[4808]: I1002 16:57:03.081434 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-ovsdbserver-sb\") pod \"dnsmasq-dns-54f9b7b8d9-fgskc\" (UID: \"9aca5301-55c6-4e78-8830-11cb79861ade\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:57:03 crc kubenswrapper[4808]: I1002 16:57:03.081555 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rbj4\" (UniqueName: \"kubernetes.io/projected/9aca5301-55c6-4e78-8830-11cb79861ade-kube-api-access-4rbj4\") pod \"dnsmasq-dns-54f9b7b8d9-fgskc\" (UID: \"9aca5301-55c6-4e78-8830-11cb79861ade\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:57:03 crc kubenswrapper[4808]: I1002 16:57:03.082088 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-config\") pod \"dnsmasq-dns-54f9b7b8d9-fgskc\" (UID: \"9aca5301-55c6-4e78-8830-11cb79861ade\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:57:03 crc kubenswrapper[4808]: I1002 16:57:03.082124 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-ovsdbserver-sb\") pod \"dnsmasq-dns-54f9b7b8d9-fgskc\" (UID: \"9aca5301-55c6-4e78-8830-11cb79861ade\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:57:03 crc kubenswrapper[4808]: I1002 16:57:03.082143 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-ovsdbserver-nb\") pod \"dnsmasq-dns-54f9b7b8d9-fgskc\" (UID: \"9aca5301-55c6-4e78-8830-11cb79861ade\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:57:03 crc kubenswrapper[4808]: I1002 16:57:03.082590 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-dns-svc\") pod \"dnsmasq-dns-54f9b7b8d9-fgskc\" (UID: \"9aca5301-55c6-4e78-8830-11cb79861ade\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:57:03 crc kubenswrapper[4808]: I1002 16:57:03.097000 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rbj4\" (UniqueName: \"kubernetes.io/projected/9aca5301-55c6-4e78-8830-11cb79861ade-kube-api-access-4rbj4\") pod \"dnsmasq-dns-54f9b7b8d9-fgskc\" (UID: \"9aca5301-55c6-4e78-8830-11cb79861ade\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:57:03 crc kubenswrapper[4808]: I1002 16:57:03.127841 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:57:03 crc kubenswrapper[4808]: I1002 16:57:03.368138 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-7wz2m" event={"ID":"cfeb1719-f48e-488b-8c22-a9be946819e3","Type":"ContainerDied","Data":"dfebc06ff291d3723b2b4d0f74c40fd34e07254748af9f3059ce479f5421f2e4"} Oct 02 16:57:03 crc kubenswrapper[4808]: I1002 16:57:03.368474 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfebc06ff291d3723b2b4d0f74c40fd34e07254748af9f3059ce479f5421f2e4" Oct 02 16:57:03 crc kubenswrapper[4808]: I1002 16:57:03.368255 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-7wz2m" Oct 02 16:57:03 crc kubenswrapper[4808]: I1002 16:57:03.603260 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-fgskc"] Oct 02 16:57:04 crc kubenswrapper[4808]: I1002 16:57:04.376385 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" event={"ID":"9aca5301-55c6-4e78-8830-11cb79861ade","Type":"ContainerStarted","Data":"6fa8caf486b55347a70536c195fc44eac7ff259ace4461681392fcd0d2879936"} Oct 02 16:57:04 crc kubenswrapper[4808]: I1002 16:57:04.378842 4808 generic.go:334] "Generic (PLEG): container finished" podID="818e02a3-653a-4b36-a68b-9d4e2df0b71e" containerID="72dfceead77ed393455b2e7fbcb7b164af0fc5382ab0ba62ad1f243b06b38645" exitCode=1 Oct 02 16:57:04 crc kubenswrapper[4808]: I1002 16:57:04.378894 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" event={"ID":"818e02a3-653a-4b36-a68b-9d4e2df0b71e","Type":"ContainerDied","Data":"72dfceead77ed393455b2e7fbcb7b164af0fc5382ab0ba62ad1f243b06b38645"} Oct 02 16:57:04 crc kubenswrapper[4808]: I1002 16:57:04.379572 4808 scope.go:117] "RemoveContainer" containerID="72dfceead77ed393455b2e7fbcb7b164af0fc5382ab0ba62ad1f243b06b38645" Oct 02 16:57:05 crc kubenswrapper[4808]: I1002 16:57:05.523172 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-4027-account-create-77w26"] Oct 02 16:57:05 crc kubenswrapper[4808]: I1002 16:57:05.524333 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4027-account-create-77w26" Oct 02 16:57:05 crc kubenswrapper[4808]: I1002 16:57:05.528126 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 02 16:57:05 crc kubenswrapper[4808]: I1002 16:57:05.534481 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-4027-account-create-77w26"] Oct 02 16:57:05 crc kubenswrapper[4808]: I1002 16:57:05.617313 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-7b56-account-create-5z4wr"] Oct 02 16:57:05 crc kubenswrapper[4808]: I1002 16:57:05.618462 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7b56-account-create-5z4wr" Oct 02 16:57:05 crc kubenswrapper[4808]: I1002 16:57:05.620859 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 02 16:57:05 crc kubenswrapper[4808]: I1002 16:57:05.630534 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grswz\" (UniqueName: \"kubernetes.io/projected/7951034c-d159-4816-8f1e-4118953ac9f0-kube-api-access-grswz\") pod \"barbican-4027-account-create-77w26\" (UID: \"7951034c-d159-4816-8f1e-4118953ac9f0\") " pod="openstack/barbican-4027-account-create-77w26" Oct 02 16:57:05 crc kubenswrapper[4808]: I1002 16:57:05.642823 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-7b56-account-create-5z4wr"] Oct 02 16:57:05 crc kubenswrapper[4808]: I1002 16:57:05.731901 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grswz\" (UniqueName: \"kubernetes.io/projected/7951034c-d159-4816-8f1e-4118953ac9f0-kube-api-access-grswz\") pod \"barbican-4027-account-create-77w26\" (UID: \"7951034c-d159-4816-8f1e-4118953ac9f0\") " pod="openstack/barbican-4027-account-create-77w26" Oct 02 16:57:05 crc kubenswrapper[4808]: I1002 16:57:05.732178 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmgd6\" (UniqueName: \"kubernetes.io/projected/f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a-kube-api-access-zmgd6\") pod \"cinder-7b56-account-create-5z4wr\" (UID: \"f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a\") " pod="openstack/cinder-7b56-account-create-5z4wr" Oct 02 16:57:05 crc kubenswrapper[4808]: I1002 16:57:05.755713 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grswz\" (UniqueName: \"kubernetes.io/projected/7951034c-d159-4816-8f1e-4118953ac9f0-kube-api-access-grswz\") pod \"barbican-4027-account-create-77w26\" (UID: \"7951034c-d159-4816-8f1e-4118953ac9f0\") " pod="openstack/barbican-4027-account-create-77w26" Oct 02 16:57:05 crc kubenswrapper[4808]: I1002 16:57:05.834344 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmgd6\" (UniqueName: \"kubernetes.io/projected/f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a-kube-api-access-zmgd6\") pod \"cinder-7b56-account-create-5z4wr\" (UID: \"f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a\") " pod="openstack/cinder-7b56-account-create-5z4wr" Oct 02 16:57:05 crc kubenswrapper[4808]: I1002 16:57:05.843568 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4027-account-create-77w26" Oct 02 16:57:05 crc kubenswrapper[4808]: I1002 16:57:05.849431 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmgd6\" (UniqueName: \"kubernetes.io/projected/f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a-kube-api-access-zmgd6\") pod \"cinder-7b56-account-create-5z4wr\" (UID: \"f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a\") " pod="openstack/cinder-7b56-account-create-5z4wr" Oct 02 16:57:05 crc kubenswrapper[4808]: I1002 16:57:05.927951 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7c90-account-create-z4xqz"] Oct 02 16:57:05 crc kubenswrapper[4808]: I1002 16:57:05.929813 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7c90-account-create-z4xqz" Oct 02 16:57:05 crc kubenswrapper[4808]: I1002 16:57:05.933269 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 02 16:57:05 crc kubenswrapper[4808]: I1002 16:57:05.934823 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7b56-account-create-5z4wr" Oct 02 16:57:05 crc kubenswrapper[4808]: I1002 16:57:05.944134 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7c90-account-create-z4xqz"] Oct 02 16:57:06 crc kubenswrapper[4808]: I1002 16:57:06.038628 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrhtr\" (UniqueName: \"kubernetes.io/projected/afc592f9-999b-4787-ae91-bdcd10aa2ad5-kube-api-access-wrhtr\") pod \"neutron-7c90-account-create-z4xqz\" (UID: \"afc592f9-999b-4787-ae91-bdcd10aa2ad5\") " pod="openstack/neutron-7c90-account-create-z4xqz" Oct 02 16:57:06 crc kubenswrapper[4808]: I1002 16:57:06.140777 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrhtr\" (UniqueName: \"kubernetes.io/projected/afc592f9-999b-4787-ae91-bdcd10aa2ad5-kube-api-access-wrhtr\") pod \"neutron-7c90-account-create-z4xqz\" (UID: \"afc592f9-999b-4787-ae91-bdcd10aa2ad5\") " pod="openstack/neutron-7c90-account-create-z4xqz" Oct 02 16:57:06 crc kubenswrapper[4808]: I1002 16:57:06.162320 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrhtr\" (UniqueName: \"kubernetes.io/projected/afc592f9-999b-4787-ae91-bdcd10aa2ad5-kube-api-access-wrhtr\") pod \"neutron-7c90-account-create-z4xqz\" (UID: \"afc592f9-999b-4787-ae91-bdcd10aa2ad5\") " pod="openstack/neutron-7c90-account-create-z4xqz" Oct 02 16:57:06 crc kubenswrapper[4808]: I1002 16:57:06.265875 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7c90-account-create-z4xqz" Oct 02 16:57:06 crc kubenswrapper[4808]: W1002 16:57:06.364636 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7951034c_d159_4816_8f1e_4118953ac9f0.slice/crio-8f77d5fed9b0b34d305293136a64b8b02918509c56efda46883697b94e7bc75a WatchSource:0}: Error finding container 8f77d5fed9b0b34d305293136a64b8b02918509c56efda46883697b94e7bc75a: Status 404 returned error can't find the container with id 8f77d5fed9b0b34d305293136a64b8b02918509c56efda46883697b94e7bc75a Oct 02 16:57:06 crc kubenswrapper[4808]: I1002 16:57:06.365077 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-4027-account-create-77w26"] Oct 02 16:57:06 crc kubenswrapper[4808]: I1002 16:57:06.419795 4808 generic.go:334] "Generic (PLEG): container finished" podID="9aca5301-55c6-4e78-8830-11cb79861ade" containerID="df1bd60d449001a0deb4d9dbe2d011019d0d5f09a9a1de3b8bbd2e2c45e06b03" exitCode=0 Oct 02 16:57:06 crc kubenswrapper[4808]: I1002 16:57:06.419881 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" event={"ID":"9aca5301-55c6-4e78-8830-11cb79861ade","Type":"ContainerDied","Data":"df1bd60d449001a0deb4d9dbe2d011019d0d5f09a9a1de3b8bbd2e2c45e06b03"} Oct 02 16:57:06 crc kubenswrapper[4808]: I1002 16:57:06.423254 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-7b56-account-create-5z4wr"] Oct 02 16:57:06 crc kubenswrapper[4808]: I1002 16:57:06.428614 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" event={"ID":"818e02a3-653a-4b36-a68b-9d4e2df0b71e","Type":"ContainerStarted","Data":"a918ee069243e5dc0b7673d53697e23876162204008e5ae0c51065fa9012a203"} Oct 02 16:57:06 crc kubenswrapper[4808]: I1002 16:57:06.429300 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 16:57:06 crc kubenswrapper[4808]: I1002 16:57:06.443841 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4027-account-create-77w26" event={"ID":"7951034c-d159-4816-8f1e-4118953ac9f0","Type":"ContainerStarted","Data":"8f77d5fed9b0b34d305293136a64b8b02918509c56efda46883697b94e7bc75a"} Oct 02 16:57:06 crc kubenswrapper[4808]: I1002 16:57:06.681373 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7c90-account-create-z4xqz"] Oct 02 16:57:07 crc kubenswrapper[4808]: I1002 16:57:07.456749 4808 generic.go:334] "Generic (PLEG): container finished" podID="7951034c-d159-4816-8f1e-4118953ac9f0" containerID="5e7f34f16e7f252280017884205c35340aad7c1c560c31c945050aff995002d2" exitCode=0 Oct 02 16:57:07 crc kubenswrapper[4808]: I1002 16:57:07.456891 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4027-account-create-77w26" event={"ID":"7951034c-d159-4816-8f1e-4118953ac9f0","Type":"ContainerDied","Data":"5e7f34f16e7f252280017884205c35340aad7c1c560c31c945050aff995002d2"} Oct 02 16:57:07 crc kubenswrapper[4808]: I1002 16:57:07.459276 4808 generic.go:334] "Generic (PLEG): container finished" podID="f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a" containerID="d94b8b03d2b3e323bd603df1e69eb862dc5f3a6f33e19b3c9b708b39f014b4d1" exitCode=0 Oct 02 16:57:07 crc kubenswrapper[4808]: I1002 16:57:07.459390 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7b56-account-create-5z4wr" event={"ID":"f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a","Type":"ContainerDied","Data":"d94b8b03d2b3e323bd603df1e69eb862dc5f3a6f33e19b3c9b708b39f014b4d1"} Oct 02 16:57:07 crc kubenswrapper[4808]: I1002 16:57:07.459439 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7b56-account-create-5z4wr" event={"ID":"f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a","Type":"ContainerStarted","Data":"5314f2af407853fd78cd5c80a908ffce510a4eb7f9fcfb1e8f737600261973f8"} Oct 02 16:57:07 crc kubenswrapper[4808]: I1002 16:57:07.462756 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" event={"ID":"9aca5301-55c6-4e78-8830-11cb79861ade","Type":"ContainerStarted","Data":"4418d941a635dd08c35aae363cd8749f5fc4adbad6b78fe5924c550b448290f0"} Oct 02 16:57:07 crc kubenswrapper[4808]: I1002 16:57:07.463646 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:57:07 crc kubenswrapper[4808]: I1002 16:57:07.464661 4808 generic.go:334] "Generic (PLEG): container finished" podID="afc592f9-999b-4787-ae91-bdcd10aa2ad5" containerID="26ed09459c52f39abfab79656cfd8c4d8c04fc13640a2314b929e472e033c6a3" exitCode=0 Oct 02 16:57:07 crc kubenswrapper[4808]: I1002 16:57:07.464750 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c90-account-create-z4xqz" event={"ID":"afc592f9-999b-4787-ae91-bdcd10aa2ad5","Type":"ContainerDied","Data":"26ed09459c52f39abfab79656cfd8c4d8c04fc13640a2314b929e472e033c6a3"} Oct 02 16:57:07 crc kubenswrapper[4808]: I1002 16:57:07.464792 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c90-account-create-z4xqz" event={"ID":"afc592f9-999b-4787-ae91-bdcd10aa2ad5","Type":"ContainerStarted","Data":"f96d7eba4edbf0afb3a23923697b5779299adb671538035126dd4c94b34e403b"} Oct 02 16:57:07 crc kubenswrapper[4808]: I1002 16:57:07.572858 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" podStartSLOduration=5.572838456 podStartE2EDuration="5.572838456s" podCreationTimestamp="2025-10-02 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:57:07.546997066 +0000 UTC m=+994.872526106" watchObservedRunningTime="2025-10-02 16:57:07.572838456 +0000 UTC m=+994.898367466" Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.019942 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7c90-account-create-z4xqz" Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.027087 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7b56-account-create-5z4wr" Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.035728 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4027-account-create-77w26" Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.086091 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmgd6\" (UniqueName: \"kubernetes.io/projected/f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a-kube-api-access-zmgd6\") pod \"f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a\" (UID: \"f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a\") " Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.086325 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrhtr\" (UniqueName: \"kubernetes.io/projected/afc592f9-999b-4787-ae91-bdcd10aa2ad5-kube-api-access-wrhtr\") pod \"afc592f9-999b-4787-ae91-bdcd10aa2ad5\" (UID: \"afc592f9-999b-4787-ae91-bdcd10aa2ad5\") " Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.092613 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a-kube-api-access-zmgd6" (OuterVolumeSpecName: "kube-api-access-zmgd6") pod "f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a" (UID: "f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a"). InnerVolumeSpecName "kube-api-access-zmgd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.092766 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afc592f9-999b-4787-ae91-bdcd10aa2ad5-kube-api-access-wrhtr" (OuterVolumeSpecName: "kube-api-access-wrhtr") pod "afc592f9-999b-4787-ae91-bdcd10aa2ad5" (UID: "afc592f9-999b-4787-ae91-bdcd10aa2ad5"). InnerVolumeSpecName "kube-api-access-wrhtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.187345 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grswz\" (UniqueName: \"kubernetes.io/projected/7951034c-d159-4816-8f1e-4118953ac9f0-kube-api-access-grswz\") pod \"7951034c-d159-4816-8f1e-4118953ac9f0\" (UID: \"7951034c-d159-4816-8f1e-4118953ac9f0\") " Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.187731 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmgd6\" (UniqueName: \"kubernetes.io/projected/f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a-kube-api-access-zmgd6\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.187749 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrhtr\" (UniqueName: \"kubernetes.io/projected/afc592f9-999b-4787-ae91-bdcd10aa2ad5-kube-api-access-wrhtr\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.191191 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7951034c-d159-4816-8f1e-4118953ac9f0-kube-api-access-grswz" (OuterVolumeSpecName: "kube-api-access-grswz") pod "7951034c-d159-4816-8f1e-4118953ac9f0" (UID: "7951034c-d159-4816-8f1e-4118953ac9f0"). InnerVolumeSpecName "kube-api-access-grswz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.289856 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grswz\" (UniqueName: \"kubernetes.io/projected/7951034c-d159-4816-8f1e-4118953ac9f0-kube-api-access-grswz\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.485955 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c90-account-create-z4xqz" event={"ID":"afc592f9-999b-4787-ae91-bdcd10aa2ad5","Type":"ContainerDied","Data":"f96d7eba4edbf0afb3a23923697b5779299adb671538035126dd4c94b34e403b"} Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.486008 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f96d7eba4edbf0afb3a23923697b5779299adb671538035126dd4c94b34e403b" Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.486067 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7c90-account-create-z4xqz" Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.492190 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4027-account-create-77w26" event={"ID":"7951034c-d159-4816-8f1e-4118953ac9f0","Type":"ContainerDied","Data":"8f77d5fed9b0b34d305293136a64b8b02918509c56efda46883697b94e7bc75a"} Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.492312 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f77d5fed9b0b34d305293136a64b8b02918509c56efda46883697b94e7bc75a" Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.492447 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4027-account-create-77w26" Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.496828 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7b56-account-create-5z4wr" Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.496720 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7b56-account-create-5z4wr" event={"ID":"f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a","Type":"ContainerDied","Data":"5314f2af407853fd78cd5c80a908ffce510a4eb7f9fcfb1e8f737600261973f8"} Oct 02 16:57:09 crc kubenswrapper[4808]: I1002 16:57:09.498135 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5314f2af407853fd78cd5c80a908ffce510a4eb7f9fcfb1e8f737600261973f8" Oct 02 16:57:10 crc kubenswrapper[4808]: I1002 16:57:10.704794 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.130561 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.205670 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-l556v"] Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.206022 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-l556v" podUID="181f7b50-6e00-4e8c-bdde-9bbff23c3306" containerName="dnsmasq-dns" containerID="cri-o://23c860288b64583e7ca3a50a120dcfd0ae8d8d7269348d26f8c69828f1139196" gracePeriod=10 Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.534960 4808 generic.go:334] "Generic (PLEG): container finished" podID="181f7b50-6e00-4e8c-bdde-9bbff23c3306" containerID="23c860288b64583e7ca3a50a120dcfd0ae8d8d7269348d26f8c69828f1139196" exitCode=0 Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.534993 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-l556v" event={"ID":"181f7b50-6e00-4e8c-bdde-9bbff23c3306","Type":"ContainerDied","Data":"23c860288b64583e7ca3a50a120dcfd0ae8d8d7269348d26f8c69828f1139196"} Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.718548 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.762756 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-config\") pod \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\" (UID: \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\") " Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.762867 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-ovsdbserver-nb\") pod \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\" (UID: \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\") " Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.762937 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-ovsdbserver-sb\") pod \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\" (UID: \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\") " Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.762968 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbvcz\" (UniqueName: \"kubernetes.io/projected/181f7b50-6e00-4e8c-bdde-9bbff23c3306-kube-api-access-sbvcz\") pod \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\" (UID: \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\") " Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.763093 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-dns-svc\") pod \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\" (UID: \"181f7b50-6e00-4e8c-bdde-9bbff23c3306\") " Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.776919 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/181f7b50-6e00-4e8c-bdde-9bbff23c3306-kube-api-access-sbvcz" (OuterVolumeSpecName: "kube-api-access-sbvcz") pod "181f7b50-6e00-4e8c-bdde-9bbff23c3306" (UID: "181f7b50-6e00-4e8c-bdde-9bbff23c3306"). InnerVolumeSpecName "kube-api-access-sbvcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.813153 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-config" (OuterVolumeSpecName: "config") pod "181f7b50-6e00-4e8c-bdde-9bbff23c3306" (UID: "181f7b50-6e00-4e8c-bdde-9bbff23c3306"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.824199 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "181f7b50-6e00-4e8c-bdde-9bbff23c3306" (UID: "181f7b50-6e00-4e8c-bdde-9bbff23c3306"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.829056 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "181f7b50-6e00-4e8c-bdde-9bbff23c3306" (UID: "181f7b50-6e00-4e8c-bdde-9bbff23c3306"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.832228 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "181f7b50-6e00-4e8c-bdde-9bbff23c3306" (UID: "181f7b50-6e00-4e8c-bdde-9bbff23c3306"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.864332 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.864361 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.864371 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.864383 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/181f7b50-6e00-4e8c-bdde-9bbff23c3306-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:13 crc kubenswrapper[4808]: I1002 16:57:13.864393 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbvcz\" (UniqueName: \"kubernetes.io/projected/181f7b50-6e00-4e8c-bdde-9bbff23c3306-kube-api-access-sbvcz\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:14 crc kubenswrapper[4808]: I1002 16:57:14.542955 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-l556v" event={"ID":"181f7b50-6e00-4e8c-bdde-9bbff23c3306","Type":"ContainerDied","Data":"d739ccd5aa66f1f2bbe82694fdbe9ed20d90d094dd768e78812394d483779901"} Oct 02 16:57:14 crc kubenswrapper[4808]: I1002 16:57:14.543023 4808 scope.go:117] "RemoveContainer" containerID="23c860288b64583e7ca3a50a120dcfd0ae8d8d7269348d26f8c69828f1139196" Oct 02 16:57:14 crc kubenswrapper[4808]: I1002 16:57:14.543052 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-l556v" Oct 02 16:57:14 crc kubenswrapper[4808]: I1002 16:57:14.569941 4808 scope.go:117] "RemoveContainer" containerID="f662f26ed8c912737d56a4c4f68e273cd086711a4650a7ae755f89b9883b82be" Oct 02 16:57:14 crc kubenswrapper[4808]: I1002 16:57:14.580384 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-l556v"] Oct 02 16:57:14 crc kubenswrapper[4808]: I1002 16:57:14.582523 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-l556v"] Oct 02 16:57:15 crc kubenswrapper[4808]: I1002 16:57:15.407551 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="181f7b50-6e00-4e8c-bdde-9bbff23c3306" path="/var/lib/kubelet/pods/181f7b50-6e00-4e8c-bdde-9bbff23c3306/volumes" Oct 02 16:57:18 crc kubenswrapper[4808]: I1002 16:57:18.739557 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 16:57:18 crc kubenswrapper[4808]: I1002 16:57:18.739835 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.063997 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-g8jl2"] Oct 02 16:57:43 crc kubenswrapper[4808]: E1002 16:57:43.065034 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc592f9-999b-4787-ae91-bdcd10aa2ad5" containerName="mariadb-account-create" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.065053 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc592f9-999b-4787-ae91-bdcd10aa2ad5" containerName="mariadb-account-create" Oct 02 16:57:43 crc kubenswrapper[4808]: E1002 16:57:43.065077 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="181f7b50-6e00-4e8c-bdde-9bbff23c3306" containerName="init" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.065086 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="181f7b50-6e00-4e8c-bdde-9bbff23c3306" containerName="init" Oct 02 16:57:43 crc kubenswrapper[4808]: E1002 16:57:43.065122 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7951034c-d159-4816-8f1e-4118953ac9f0" containerName="mariadb-account-create" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.065132 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7951034c-d159-4816-8f1e-4118953ac9f0" containerName="mariadb-account-create" Oct 02 16:57:43 crc kubenswrapper[4808]: E1002 16:57:43.065147 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="181f7b50-6e00-4e8c-bdde-9bbff23c3306" containerName="dnsmasq-dns" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.065155 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="181f7b50-6e00-4e8c-bdde-9bbff23c3306" containerName="dnsmasq-dns" Oct 02 16:57:43 crc kubenswrapper[4808]: E1002 16:57:43.065170 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a" containerName="mariadb-account-create" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.065178 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a" containerName="mariadb-account-create" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.065380 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7951034c-d159-4816-8f1e-4118953ac9f0" containerName="mariadb-account-create" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.065404 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="afc592f9-999b-4787-ae91-bdcd10aa2ad5" containerName="mariadb-account-create" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.065413 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a" containerName="mariadb-account-create" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.065428 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="181f7b50-6e00-4e8c-bdde-9bbff23c3306" containerName="dnsmasq-dns" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.066052 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.068902 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.069402 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-b9ff9" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.069627 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.073447 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.078389 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-ktb4l"] Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.079653 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.098050 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-ktb4l"] Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.114412 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-g8jl2"] Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.173383 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-combined-ca-bundle\") pod \"keystone-bootstrap-g8jl2\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.174385 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-fernet-keys\") pod \"keystone-bootstrap-g8jl2\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.174447 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-config-data\") pod \"keystone-bootstrap-g8jl2\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.174475 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-scripts\") pod \"keystone-bootstrap-g8jl2\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.174517 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvkks\" (UniqueName: \"kubernetes.io/projected/e976f05e-37b0-4803-bff2-b7b164c0e24d-kube-api-access-nvkks\") pod \"keystone-bootstrap-g8jl2\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.174553 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-credential-keys\") pod \"keystone-bootstrap-g8jl2\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.281033 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-config-data\") pod \"keystone-bootstrap-g8jl2\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.281089 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-scripts\") pod \"keystone-bootstrap-g8jl2\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.281111 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-ovsdbserver-sb\") pod \"dnsmasq-dns-6546db6db7-ktb4l\" (UID: \"7cd07478-020c-4aee-b27f-9a193440b09e\") " pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.281144 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhkzw\" (UniqueName: \"kubernetes.io/projected/7cd07478-020c-4aee-b27f-9a193440b09e-kube-api-access-xhkzw\") pod \"dnsmasq-dns-6546db6db7-ktb4l\" (UID: \"7cd07478-020c-4aee-b27f-9a193440b09e\") " pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.281169 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvkks\" (UniqueName: \"kubernetes.io/projected/e976f05e-37b0-4803-bff2-b7b164c0e24d-kube-api-access-nvkks\") pod \"keystone-bootstrap-g8jl2\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.281194 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-ovsdbserver-nb\") pod \"dnsmasq-dns-6546db6db7-ktb4l\" (UID: \"7cd07478-020c-4aee-b27f-9a193440b09e\") " pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.281214 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-credential-keys\") pod \"keystone-bootstrap-g8jl2\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.281280 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-config\") pod \"dnsmasq-dns-6546db6db7-ktb4l\" (UID: \"7cd07478-020c-4aee-b27f-9a193440b09e\") " pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.281301 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-dns-svc\") pod \"dnsmasq-dns-6546db6db7-ktb4l\" (UID: \"7cd07478-020c-4aee-b27f-9a193440b09e\") " pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.281337 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-combined-ca-bundle\") pod \"keystone-bootstrap-g8jl2\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.281359 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-fernet-keys\") pod \"keystone-bootstrap-g8jl2\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.292063 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-config-data\") pod \"keystone-bootstrap-g8jl2\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.292585 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-combined-ca-bundle\") pod \"keystone-bootstrap-g8jl2\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.303638 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-scripts\") pod \"keystone-bootstrap-g8jl2\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.303962 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-credential-keys\") pod \"keystone-bootstrap-g8jl2\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.304119 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-fernet-keys\") pod \"keystone-bootstrap-g8jl2\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.310826 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvkks\" (UniqueName: \"kubernetes.io/projected/e976f05e-37b0-4803-bff2-b7b164c0e24d-kube-api-access-nvkks\") pod \"keystone-bootstrap-g8jl2\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.351360 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-l2fls"] Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.352676 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-l2fls" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.363642 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.383514 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-config\") pod \"dnsmasq-dns-6546db6db7-ktb4l\" (UID: \"7cd07478-020c-4aee-b27f-9a193440b09e\") " pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.383559 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-dns-svc\") pod \"dnsmasq-dns-6546db6db7-ktb4l\" (UID: \"7cd07478-020c-4aee-b27f-9a193440b09e\") " pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.383583 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-combined-ca-bundle\") pod \"cinder-db-sync-l2fls\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " pod="openstack/cinder-db-sync-l2fls" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.383610 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd29c\" (UniqueName: \"kubernetes.io/projected/b3d4166a-c265-4cb4-8481-32e6b0144997-kube-api-access-bd29c\") pod \"cinder-db-sync-l2fls\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " pod="openstack/cinder-db-sync-l2fls" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.383662 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-scripts\") pod \"cinder-db-sync-l2fls\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " pod="openstack/cinder-db-sync-l2fls" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.383683 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-config-data\") pod \"cinder-db-sync-l2fls\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " pod="openstack/cinder-db-sync-l2fls" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.383708 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-ovsdbserver-sb\") pod \"dnsmasq-dns-6546db6db7-ktb4l\" (UID: \"7cd07478-020c-4aee-b27f-9a193440b09e\") " pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.383733 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhkzw\" (UniqueName: \"kubernetes.io/projected/7cd07478-020c-4aee-b27f-9a193440b09e-kube-api-access-xhkzw\") pod \"dnsmasq-dns-6546db6db7-ktb4l\" (UID: \"7cd07478-020c-4aee-b27f-9a193440b09e\") " pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.383752 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-db-sync-config-data\") pod \"cinder-db-sync-l2fls\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " pod="openstack/cinder-db-sync-l2fls" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.383779 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-ovsdbserver-nb\") pod \"dnsmasq-dns-6546db6db7-ktb4l\" (UID: \"7cd07478-020c-4aee-b27f-9a193440b09e\") " pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.383848 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b3d4166a-c265-4cb4-8481-32e6b0144997-etc-machine-id\") pod \"cinder-db-sync-l2fls\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " pod="openstack/cinder-db-sync-l2fls" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.384152 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-q2dqb" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.384888 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-dns-svc\") pod \"dnsmasq-dns-6546db6db7-ktb4l\" (UID: \"7cd07478-020c-4aee-b27f-9a193440b09e\") " pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.384962 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-config\") pod \"dnsmasq-dns-6546db6db7-ktb4l\" (UID: \"7cd07478-020c-4aee-b27f-9a193440b09e\") " pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.385015 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-ovsdbserver-sb\") pod \"dnsmasq-dns-6546db6db7-ktb4l\" (UID: \"7cd07478-020c-4aee-b27f-9a193440b09e\") " pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.385254 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-ovsdbserver-nb\") pod \"dnsmasq-dns-6546db6db7-ktb4l\" (UID: \"7cd07478-020c-4aee-b27f-9a193440b09e\") " pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.387595 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.399534 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.405592 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-l2fls"] Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.447919 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhkzw\" (UniqueName: \"kubernetes.io/projected/7cd07478-020c-4aee-b27f-9a193440b09e-kube-api-access-xhkzw\") pod \"dnsmasq-dns-6546db6db7-ktb4l\" (UID: \"7cd07478-020c-4aee-b27f-9a193440b09e\") " pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.451452 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-dl6kd"] Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.468998 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dl6kd" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.492318 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-db-sync-config-data\") pod \"cinder-db-sync-l2fls\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " pod="openstack/cinder-db-sync-l2fls" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.492437 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b3d4166a-c265-4cb4-8481-32e6b0144997-etc-machine-id\") pod \"cinder-db-sync-l2fls\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " pod="openstack/cinder-db-sync-l2fls" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.492484 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-combined-ca-bundle\") pod \"cinder-db-sync-l2fls\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " pod="openstack/cinder-db-sync-l2fls" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.492529 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd29c\" (UniqueName: \"kubernetes.io/projected/b3d4166a-c265-4cb4-8481-32e6b0144997-kube-api-access-bd29c\") pod \"cinder-db-sync-l2fls\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " pod="openstack/cinder-db-sync-l2fls" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.492614 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-scripts\") pod \"cinder-db-sync-l2fls\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " pod="openstack/cinder-db-sync-l2fls" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.492632 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-config-data\") pod \"cinder-db-sync-l2fls\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " pod="openstack/cinder-db-sync-l2fls" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.494062 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b3d4166a-c265-4cb4-8481-32e6b0144997-etc-machine-id\") pod \"cinder-db-sync-l2fls\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " pod="openstack/cinder-db-sync-l2fls" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.499526 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-dl6kd"] Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.499586 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-ckf4j"] Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.518672 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-scripts\") pod \"cinder-db-sync-l2fls\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " pod="openstack/cinder-db-sync-l2fls" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.519428 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-combined-ca-bundle\") pod \"cinder-db-sync-l2fls\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " pod="openstack/cinder-db-sync-l2fls" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.520423 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-db-sync-config-data\") pod \"cinder-db-sync-l2fls\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " pod="openstack/cinder-db-sync-l2fls" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.527374 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ckf4j" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.535386 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-config-data\") pod \"cinder-db-sync-l2fls\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " pod="openstack/cinder-db-sync-l2fls" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.536926 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.537184 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.537452 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-wbbc4" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.537563 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-mfvng" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.537676 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.540552 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd29c\" (UniqueName: \"kubernetes.io/projected/b3d4166a-c265-4cb4-8481-32e6b0144997-kube-api-access-bd29c\") pod \"cinder-db-sync-l2fls\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " pod="openstack/cinder-db-sync-l2fls" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.570819 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ckf4j"] Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.601096 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a437827-755a-4648-a40e-18a8ef89355d-logs\") pod \"placement-db-sync-dl6kd\" (UID: \"7a437827-755a-4648-a40e-18a8ef89355d\") " pod="openstack/placement-db-sync-dl6kd" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.601159 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a437827-755a-4648-a40e-18a8ef89355d-combined-ca-bundle\") pod \"placement-db-sync-dl6kd\" (UID: \"7a437827-755a-4648-a40e-18a8ef89355d\") " pod="openstack/placement-db-sync-dl6kd" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.601220 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a437827-755a-4648-a40e-18a8ef89355d-config-data\") pod \"placement-db-sync-dl6kd\" (UID: \"7a437827-755a-4648-a40e-18a8ef89355d\") " pod="openstack/placement-db-sync-dl6kd" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.612322 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a437827-755a-4648-a40e-18a8ef89355d-scripts\") pod \"placement-db-sync-dl6kd\" (UID: \"7a437827-755a-4648-a40e-18a8ef89355d\") " pod="openstack/placement-db-sync-dl6kd" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.612728 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfntz\" (UniqueName: \"kubernetes.io/projected/7a437827-755a-4648-a40e-18a8ef89355d-kube-api-access-wfntz\") pod \"placement-db-sync-dl6kd\" (UID: \"7a437827-755a-4648-a40e-18a8ef89355d\") " pod="openstack/placement-db-sync-dl6kd" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.642477 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-ktb4l"] Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.643608 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.673183 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.675457 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.678562 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.678576 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.686872 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-l2fls" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.692498 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.700153 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-kkks2"] Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.701405 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.708361 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-kkks2"] Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.718827 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a1c6ad-0412-4c46-90ce-34869c9ca9e6-combined-ca-bundle\") pod \"barbican-db-sync-ckf4j\" (UID: \"74a1c6ad-0412-4c46-90ce-34869c9ca9e6\") " pod="openstack/barbican-db-sync-ckf4j" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.718891 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a437827-755a-4648-a40e-18a8ef89355d-config-data\") pod \"placement-db-sync-dl6kd\" (UID: \"7a437827-755a-4648-a40e-18a8ef89355d\") " pod="openstack/placement-db-sync-dl6kd" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.718934 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a437827-755a-4648-a40e-18a8ef89355d-scripts\") pod \"placement-db-sync-dl6kd\" (UID: \"7a437827-755a-4648-a40e-18a8ef89355d\") " pod="openstack/placement-db-sync-dl6kd" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.719027 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhth4\" (UniqueName: \"kubernetes.io/projected/74a1c6ad-0412-4c46-90ce-34869c9ca9e6-kube-api-access-jhth4\") pod \"barbican-db-sync-ckf4j\" (UID: \"74a1c6ad-0412-4c46-90ce-34869c9ca9e6\") " pod="openstack/barbican-db-sync-ckf4j" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.719066 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/74a1c6ad-0412-4c46-90ce-34869c9ca9e6-db-sync-config-data\") pod \"barbican-db-sync-ckf4j\" (UID: \"74a1c6ad-0412-4c46-90ce-34869c9ca9e6\") " pod="openstack/barbican-db-sync-ckf4j" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.719101 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfntz\" (UniqueName: \"kubernetes.io/projected/7a437827-755a-4648-a40e-18a8ef89355d-kube-api-access-wfntz\") pod \"placement-db-sync-dl6kd\" (UID: \"7a437827-755a-4648-a40e-18a8ef89355d\") " pod="openstack/placement-db-sync-dl6kd" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.719129 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a437827-755a-4648-a40e-18a8ef89355d-logs\") pod \"placement-db-sync-dl6kd\" (UID: \"7a437827-755a-4648-a40e-18a8ef89355d\") " pod="openstack/placement-db-sync-dl6kd" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.719190 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a437827-755a-4648-a40e-18a8ef89355d-combined-ca-bundle\") pod \"placement-db-sync-dl6kd\" (UID: \"7a437827-755a-4648-a40e-18a8ef89355d\") " pod="openstack/placement-db-sync-dl6kd" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.721973 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a437827-755a-4648-a40e-18a8ef89355d-logs\") pod \"placement-db-sync-dl6kd\" (UID: \"7a437827-755a-4648-a40e-18a8ef89355d\") " pod="openstack/placement-db-sync-dl6kd" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.731490 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-zpkmw"] Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.732663 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-zpkmw" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.734312 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a437827-755a-4648-a40e-18a8ef89355d-combined-ca-bundle\") pod \"placement-db-sync-dl6kd\" (UID: \"7a437827-755a-4648-a40e-18a8ef89355d\") " pod="openstack/placement-db-sync-dl6kd" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.734892 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.735107 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-cprk7" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.737573 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.739339 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a437827-755a-4648-a40e-18a8ef89355d-scripts\") pod \"placement-db-sync-dl6kd\" (UID: \"7a437827-755a-4648-a40e-18a8ef89355d\") " pod="openstack/placement-db-sync-dl6kd" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.739891 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a437827-755a-4648-a40e-18a8ef89355d-config-data\") pod \"placement-db-sync-dl6kd\" (UID: \"7a437827-755a-4648-a40e-18a8ef89355d\") " pod="openstack/placement-db-sync-dl6kd" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.750444 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-zpkmw"] Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.757737 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfntz\" (UniqueName: \"kubernetes.io/projected/7a437827-755a-4648-a40e-18a8ef89355d-kube-api-access-wfntz\") pod \"placement-db-sync-dl6kd\" (UID: \"7a437827-755a-4648-a40e-18a8ef89355d\") " pod="openstack/placement-db-sync-dl6kd" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.821037 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/435bbb18-4b67-4b64-a922-f57717d8adc8-run-httpd\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.821095 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-ovsdbserver-sb\") pod \"dnsmasq-dns-7987f74bbc-kkks2\" (UID: \"92a07292-975e-492a-97da-f713cfb5dbce\") " pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.821153 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.821173 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-ovsdbserver-nb\") pod \"dnsmasq-dns-7987f74bbc-kkks2\" (UID: \"92a07292-975e-492a-97da-f713cfb5dbce\") " pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.821203 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a1c6ad-0412-4c46-90ce-34869c9ca9e6-combined-ca-bundle\") pod \"barbican-db-sync-ckf4j\" (UID: \"74a1c6ad-0412-4c46-90ce-34869c9ca9e6\") " pod="openstack/barbican-db-sync-ckf4j" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.821243 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-dns-svc\") pod \"dnsmasq-dns-7987f74bbc-kkks2\" (UID: \"92a07292-975e-492a-97da-f713cfb5dbce\") " pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.821268 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.821291 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mw2j\" (UniqueName: \"kubernetes.io/projected/92a07292-975e-492a-97da-f713cfb5dbce-kube-api-access-4mw2j\") pod \"dnsmasq-dns-7987f74bbc-kkks2\" (UID: \"92a07292-975e-492a-97da-f713cfb5dbce\") " pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.821312 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-config\") pod \"dnsmasq-dns-7987f74bbc-kkks2\" (UID: \"92a07292-975e-492a-97da-f713cfb5dbce\") " pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.821331 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/435bbb18-4b67-4b64-a922-f57717d8adc8-log-httpd\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.821354 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-scripts\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.821384 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-config-data\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.821413 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhth4\" (UniqueName: \"kubernetes.io/projected/74a1c6ad-0412-4c46-90ce-34869c9ca9e6-kube-api-access-jhth4\") pod \"barbican-db-sync-ckf4j\" (UID: \"74a1c6ad-0412-4c46-90ce-34869c9ca9e6\") " pod="openstack/barbican-db-sync-ckf4j" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.821437 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtfq8\" (UniqueName: \"kubernetes.io/projected/435bbb18-4b67-4b64-a922-f57717d8adc8-kube-api-access-xtfq8\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.821456 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/74a1c6ad-0412-4c46-90ce-34869c9ca9e6-db-sync-config-data\") pod \"barbican-db-sync-ckf4j\" (UID: \"74a1c6ad-0412-4c46-90ce-34869c9ca9e6\") " pod="openstack/barbican-db-sync-ckf4j" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.827134 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/74a1c6ad-0412-4c46-90ce-34869c9ca9e6-db-sync-config-data\") pod \"barbican-db-sync-ckf4j\" (UID: \"74a1c6ad-0412-4c46-90ce-34869c9ca9e6\") " pod="openstack/barbican-db-sync-ckf4j" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.829428 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a1c6ad-0412-4c46-90ce-34869c9ca9e6-combined-ca-bundle\") pod \"barbican-db-sync-ckf4j\" (UID: \"74a1c6ad-0412-4c46-90ce-34869c9ca9e6\") " pod="openstack/barbican-db-sync-ckf4j" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.844474 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhth4\" (UniqueName: \"kubernetes.io/projected/74a1c6ad-0412-4c46-90ce-34869c9ca9e6-kube-api-access-jhth4\") pod \"barbican-db-sync-ckf4j\" (UID: \"74a1c6ad-0412-4c46-90ce-34869c9ca9e6\") " pod="openstack/barbican-db-sync-ckf4j" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.923278 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtfq8\" (UniqueName: \"kubernetes.io/projected/435bbb18-4b67-4b64-a922-f57717d8adc8-kube-api-access-xtfq8\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.923354 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c50129b6-292e-434b-a617-2781dac4101e-config\") pod \"neutron-db-sync-zpkmw\" (UID: \"c50129b6-292e-434b-a617-2781dac4101e\") " pod="openstack/neutron-db-sync-zpkmw" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.923375 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/435bbb18-4b67-4b64-a922-f57717d8adc8-run-httpd\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.923414 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-ovsdbserver-sb\") pod \"dnsmasq-dns-7987f74bbc-kkks2\" (UID: \"92a07292-975e-492a-97da-f713cfb5dbce\") " pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.923536 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.923587 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-ovsdbserver-nb\") pod \"dnsmasq-dns-7987f74bbc-kkks2\" (UID: \"92a07292-975e-492a-97da-f713cfb5dbce\") " pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.923652 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-dns-svc\") pod \"dnsmasq-dns-7987f74bbc-kkks2\" (UID: \"92a07292-975e-492a-97da-f713cfb5dbce\") " pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.923687 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.923718 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mw2j\" (UniqueName: \"kubernetes.io/projected/92a07292-975e-492a-97da-f713cfb5dbce-kube-api-access-4mw2j\") pod \"dnsmasq-dns-7987f74bbc-kkks2\" (UID: \"92a07292-975e-492a-97da-f713cfb5dbce\") " pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.923742 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-config\") pod \"dnsmasq-dns-7987f74bbc-kkks2\" (UID: \"92a07292-975e-492a-97da-f713cfb5dbce\") " pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.923765 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/435bbb18-4b67-4b64-a922-f57717d8adc8-log-httpd\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.923797 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c50129b6-292e-434b-a617-2781dac4101e-combined-ca-bundle\") pod \"neutron-db-sync-zpkmw\" (UID: \"c50129b6-292e-434b-a617-2781dac4101e\") " pod="openstack/neutron-db-sync-zpkmw" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.923825 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-scripts\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.923883 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-config-data\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.923903 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcqrq\" (UniqueName: \"kubernetes.io/projected/c50129b6-292e-434b-a617-2781dac4101e-kube-api-access-qcqrq\") pod \"neutron-db-sync-zpkmw\" (UID: \"c50129b6-292e-434b-a617-2781dac4101e\") " pod="openstack/neutron-db-sync-zpkmw" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.923822 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/435bbb18-4b67-4b64-a922-f57717d8adc8-run-httpd\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.924847 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-config\") pod \"dnsmasq-dns-7987f74bbc-kkks2\" (UID: \"92a07292-975e-492a-97da-f713cfb5dbce\") " pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.924910 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-ovsdbserver-nb\") pod \"dnsmasq-dns-7987f74bbc-kkks2\" (UID: \"92a07292-975e-492a-97da-f713cfb5dbce\") " pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.924944 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-dns-svc\") pod \"dnsmasq-dns-7987f74bbc-kkks2\" (UID: \"92a07292-975e-492a-97da-f713cfb5dbce\") " pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.924959 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/435bbb18-4b67-4b64-a922-f57717d8adc8-log-httpd\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.928364 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.928595 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-scripts\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.927972 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-ovsdbserver-sb\") pod \"dnsmasq-dns-7987f74bbc-kkks2\" (UID: \"92a07292-975e-492a-97da-f713cfb5dbce\") " pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.933764 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.937472 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-config-data\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.941804 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtfq8\" (UniqueName: \"kubernetes.io/projected/435bbb18-4b67-4b64-a922-f57717d8adc8-kube-api-access-xtfq8\") pod \"ceilometer-0\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " pod="openstack/ceilometer-0" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.941965 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mw2j\" (UniqueName: \"kubernetes.io/projected/92a07292-975e-492a-97da-f713cfb5dbce-kube-api-access-4mw2j\") pod \"dnsmasq-dns-7987f74bbc-kkks2\" (UID: \"92a07292-975e-492a-97da-f713cfb5dbce\") " pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.962756 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dl6kd" Oct 02 16:57:43 crc kubenswrapper[4808]: I1002 16:57:43.999385 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ckf4j" Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.026364 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c50129b6-292e-434b-a617-2781dac4101e-combined-ca-bundle\") pod \"neutron-db-sync-zpkmw\" (UID: \"c50129b6-292e-434b-a617-2781dac4101e\") " pod="openstack/neutron-db-sync-zpkmw" Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.026460 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcqrq\" (UniqueName: \"kubernetes.io/projected/c50129b6-292e-434b-a617-2781dac4101e-kube-api-access-qcqrq\") pod \"neutron-db-sync-zpkmw\" (UID: \"c50129b6-292e-434b-a617-2781dac4101e\") " pod="openstack/neutron-db-sync-zpkmw" Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.026545 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c50129b6-292e-434b-a617-2781dac4101e-config\") pod \"neutron-db-sync-zpkmw\" (UID: \"c50129b6-292e-434b-a617-2781dac4101e\") " pod="openstack/neutron-db-sync-zpkmw" Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.034922 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c50129b6-292e-434b-a617-2781dac4101e-config\") pod \"neutron-db-sync-zpkmw\" (UID: \"c50129b6-292e-434b-a617-2781dac4101e\") " pod="openstack/neutron-db-sync-zpkmw" Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.035410 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c50129b6-292e-434b-a617-2781dac4101e-combined-ca-bundle\") pod \"neutron-db-sync-zpkmw\" (UID: \"c50129b6-292e-434b-a617-2781dac4101e\") " pod="openstack/neutron-db-sync-zpkmw" Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.037597 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.045163 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcqrq\" (UniqueName: \"kubernetes.io/projected/c50129b6-292e-434b-a617-2781dac4101e-kube-api-access-qcqrq\") pod \"neutron-db-sync-zpkmw\" (UID: \"c50129b6-292e-434b-a617-2781dac4101e\") " pod="openstack/neutron-db-sync-zpkmw" Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.046608 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-g8jl2"] Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.058543 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:57:44 crc kubenswrapper[4808]: W1002 16:57:44.060843 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode976f05e_37b0_4803_bff2_b7b164c0e24d.slice/crio-500c8bb33f0ef07fd09e291fdcecbd601e1449afb85599056aa8a3c69d8659f2 WatchSource:0}: Error finding container 500c8bb33f0ef07fd09e291fdcecbd601e1449afb85599056aa8a3c69d8659f2: Status 404 returned error can't find the container with id 500c8bb33f0ef07fd09e291fdcecbd601e1449afb85599056aa8a3c69d8659f2 Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.068347 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-zpkmw" Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.194846 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-l2fls"] Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.222843 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-ktb4l"] Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.427649 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-dl6kd"] Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.576818 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.603402 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ckf4j"] Oct 02 16:57:44 crc kubenswrapper[4808]: W1002 16:57:44.609475 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74a1c6ad_0412_4c46_90ce_34869c9ca9e6.slice/crio-9ad521c143efbef049450fa0e7f2ea88874eb7a29c46eeb91bbeca7ca65cb823 WatchSource:0}: Error finding container 9ad521c143efbef049450fa0e7f2ea88874eb7a29c46eeb91bbeca7ca65cb823: Status 404 returned error can't find the container with id 9ad521c143efbef049450fa0e7f2ea88874eb7a29c46eeb91bbeca7ca65cb823 Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.694620 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-zpkmw"] Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.709475 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-kkks2"] Oct 02 16:57:44 crc kubenswrapper[4808]: W1002 16:57:44.719818 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc50129b6_292e_434b_a617_2781dac4101e.slice/crio-0634d96b8fdf947f50283e7c7b55a7becfb92b15117cbc3c3c6d7ec534ece3e4 WatchSource:0}: Error finding container 0634d96b8fdf947f50283e7c7b55a7becfb92b15117cbc3c3c6d7ec534ece3e4: Status 404 returned error can't find the container with id 0634d96b8fdf947f50283e7c7b55a7becfb92b15117cbc3c3c6d7ec534ece3e4 Oct 02 16:57:44 crc kubenswrapper[4808]: W1002 16:57:44.720197 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92a07292_975e_492a_97da_f713cfb5dbce.slice/crio-1d9f0e18f3c3062056a3554266354abe51a2816235d434999066da7714fd2efd WatchSource:0}: Error finding container 1d9f0e18f3c3062056a3554266354abe51a2816235d434999066da7714fd2efd: Status 404 returned error can't find the container with id 1d9f0e18f3c3062056a3554266354abe51a2816235d434999066da7714fd2efd Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.851464 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ckf4j" event={"ID":"74a1c6ad-0412-4c46-90ce-34869c9ca9e6","Type":"ContainerStarted","Data":"9ad521c143efbef049450fa0e7f2ea88874eb7a29c46eeb91bbeca7ca65cb823"} Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.852435 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-l2fls" event={"ID":"b3d4166a-c265-4cb4-8481-32e6b0144997","Type":"ContainerStarted","Data":"a955a03dc136394e57f80c0c6eeeb7b7a798bef982992e563dfae51f03c36bfe"} Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.857773 4808 generic.go:334] "Generic (PLEG): container finished" podID="7cd07478-020c-4aee-b27f-9a193440b09e" containerID="7c886459d2c33f92b42fff25a9011562de758449bd28896e4db9009e612c0ca1" exitCode=0 Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.857842 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" event={"ID":"7cd07478-020c-4aee-b27f-9a193440b09e","Type":"ContainerDied","Data":"7c886459d2c33f92b42fff25a9011562de758449bd28896e4db9009e612c0ca1"} Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.857868 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" event={"ID":"7cd07478-020c-4aee-b27f-9a193440b09e","Type":"ContainerStarted","Data":"ef0c674056fc3103d30906354d8aedc9957078b1c0c3ef8669a971873a7440bb"} Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.860282 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"435bbb18-4b67-4b64-a922-f57717d8adc8","Type":"ContainerStarted","Data":"57a9233bdfbfe05a8a59069e6e6442792909c1c901539f75f70adf637b12f350"} Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.863560 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-g8jl2" event={"ID":"e976f05e-37b0-4803-bff2-b7b164c0e24d","Type":"ContainerStarted","Data":"28e4a8f6f8eaffa1abe22e2ae1b0be1e3da2c9034034dc4b4e8895d8a6c51d1b"} Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.863592 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-g8jl2" event={"ID":"e976f05e-37b0-4803-bff2-b7b164c0e24d","Type":"ContainerStarted","Data":"500c8bb33f0ef07fd09e291fdcecbd601e1449afb85599056aa8a3c69d8659f2"} Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.866731 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dl6kd" event={"ID":"7a437827-755a-4648-a40e-18a8ef89355d","Type":"ContainerStarted","Data":"fc7c521aba60bd23b39c3b2578eae342f5f48e3148fabfc107e086fe1104a1ad"} Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.868282 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-zpkmw" event={"ID":"c50129b6-292e-434b-a617-2781dac4101e","Type":"ContainerStarted","Data":"0634d96b8fdf947f50283e7c7b55a7becfb92b15117cbc3c3c6d7ec534ece3e4"} Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.874534 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" event={"ID":"92a07292-975e-492a-97da-f713cfb5dbce","Type":"ContainerStarted","Data":"1d9f0e18f3c3062056a3554266354abe51a2816235d434999066da7714fd2efd"} Oct 02 16:57:44 crc kubenswrapper[4808]: I1002 16:57:44.911326 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-g8jl2" podStartSLOduration=1.911308563 podStartE2EDuration="1.911308563s" podCreationTimestamp="2025-10-02 16:57:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:57:44.906687808 +0000 UTC m=+1032.232216808" watchObservedRunningTime="2025-10-02 16:57:44.911308563 +0000 UTC m=+1032.236837563" Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.174661 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.356216 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-dns-svc\") pod \"7cd07478-020c-4aee-b27f-9a193440b09e\" (UID: \"7cd07478-020c-4aee-b27f-9a193440b09e\") " Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.356302 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-ovsdbserver-nb\") pod \"7cd07478-020c-4aee-b27f-9a193440b09e\" (UID: \"7cd07478-020c-4aee-b27f-9a193440b09e\") " Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.356464 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-ovsdbserver-sb\") pod \"7cd07478-020c-4aee-b27f-9a193440b09e\" (UID: \"7cd07478-020c-4aee-b27f-9a193440b09e\") " Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.356482 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-config\") pod \"7cd07478-020c-4aee-b27f-9a193440b09e\" (UID: \"7cd07478-020c-4aee-b27f-9a193440b09e\") " Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.356526 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhkzw\" (UniqueName: \"kubernetes.io/projected/7cd07478-020c-4aee-b27f-9a193440b09e-kube-api-access-xhkzw\") pod \"7cd07478-020c-4aee-b27f-9a193440b09e\" (UID: \"7cd07478-020c-4aee-b27f-9a193440b09e\") " Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.360513 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cd07478-020c-4aee-b27f-9a193440b09e-kube-api-access-xhkzw" (OuterVolumeSpecName: "kube-api-access-xhkzw") pod "7cd07478-020c-4aee-b27f-9a193440b09e" (UID: "7cd07478-020c-4aee-b27f-9a193440b09e"). InnerVolumeSpecName "kube-api-access-xhkzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.388117 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7cd07478-020c-4aee-b27f-9a193440b09e" (UID: "7cd07478-020c-4aee-b27f-9a193440b09e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.394058 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7cd07478-020c-4aee-b27f-9a193440b09e" (UID: "7cd07478-020c-4aee-b27f-9a193440b09e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.398362 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-config" (OuterVolumeSpecName: "config") pod "7cd07478-020c-4aee-b27f-9a193440b09e" (UID: "7cd07478-020c-4aee-b27f-9a193440b09e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.435703 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7cd07478-020c-4aee-b27f-9a193440b09e" (UID: "7cd07478-020c-4aee-b27f-9a193440b09e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.459691 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.459726 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.459738 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhkzw\" (UniqueName: \"kubernetes.io/projected/7cd07478-020c-4aee-b27f-9a193440b09e-kube-api-access-xhkzw\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.459749 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.459757 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cd07478-020c-4aee-b27f-9a193440b09e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.594466 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.904026 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" event={"ID":"7cd07478-020c-4aee-b27f-9a193440b09e","Type":"ContainerDied","Data":"ef0c674056fc3103d30906354d8aedc9957078b1c0c3ef8669a971873a7440bb"} Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.904130 4808 scope.go:117] "RemoveContainer" containerID="7c886459d2c33f92b42fff25a9011562de758449bd28896e4db9009e612c0ca1" Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.904347 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6546db6db7-ktb4l" Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.915304 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-zpkmw" event={"ID":"c50129b6-292e-434b-a617-2781dac4101e","Type":"ContainerStarted","Data":"e11d95d6c344e8547933c8bf98688be3e6c5e1d7a02e78850fed4d4acf0bf0ed"} Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.918644 4808 generic.go:334] "Generic (PLEG): container finished" podID="92a07292-975e-492a-97da-f713cfb5dbce" containerID="eec0e4699b3d8afef7c91bba2e79f85f1a0d2dd638a7df19ddbf3a42b2d33619" exitCode=0 Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.919678 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" event={"ID":"92a07292-975e-492a-97da-f713cfb5dbce","Type":"ContainerDied","Data":"eec0e4699b3d8afef7c91bba2e79f85f1a0d2dd638a7df19ddbf3a42b2d33619"} Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.934095 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-zpkmw" podStartSLOduration=2.934077335 podStartE2EDuration="2.934077335s" podCreationTimestamp="2025-10-02 16:57:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:57:45.927630191 +0000 UTC m=+1033.253159191" watchObservedRunningTime="2025-10-02 16:57:45.934077335 +0000 UTC m=+1033.259606335" Oct 02 16:57:45 crc kubenswrapper[4808]: I1002 16:57:45.999733 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-ktb4l"] Oct 02 16:57:46 crc kubenswrapper[4808]: I1002 16:57:46.011927 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-ktb4l"] Oct 02 16:57:46 crc kubenswrapper[4808]: I1002 16:57:46.934199 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" event={"ID":"92a07292-975e-492a-97da-f713cfb5dbce","Type":"ContainerStarted","Data":"b398c4c8f8e515341ac81d5c11e1b88348a2cff349306d7299c715065f1530cb"} Oct 02 16:57:46 crc kubenswrapper[4808]: I1002 16:57:46.934650 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:57:46 crc kubenswrapper[4808]: I1002 16:57:46.955829 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" podStartSLOduration=3.955806821 podStartE2EDuration="3.955806821s" podCreationTimestamp="2025-10-02 16:57:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:57:46.949764697 +0000 UTC m=+1034.275293697" watchObservedRunningTime="2025-10-02 16:57:46.955806821 +0000 UTC m=+1034.281335831" Oct 02 16:57:47 crc kubenswrapper[4808]: I1002 16:57:47.406181 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cd07478-020c-4aee-b27f-9a193440b09e" path="/var/lib/kubelet/pods/7cd07478-020c-4aee-b27f-9a193440b09e/volumes" Oct 02 16:57:48 crc kubenswrapper[4808]: I1002 16:57:48.739426 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 16:57:48 crc kubenswrapper[4808]: I1002 16:57:48.739738 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 16:57:48 crc kubenswrapper[4808]: I1002 16:57:48.952653 4808 generic.go:334] "Generic (PLEG): container finished" podID="e976f05e-37b0-4803-bff2-b7b164c0e24d" containerID="28e4a8f6f8eaffa1abe22e2ae1b0be1e3da2c9034034dc4b4e8895d8a6c51d1b" exitCode=0 Oct 02 16:57:48 crc kubenswrapper[4808]: I1002 16:57:48.952720 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-g8jl2" event={"ID":"e976f05e-37b0-4803-bff2-b7b164c0e24d","Type":"ContainerDied","Data":"28e4a8f6f8eaffa1abe22e2ae1b0be1e3da2c9034034dc4b4e8895d8a6c51d1b"} Oct 02 16:57:53 crc kubenswrapper[4808]: I1002 16:57:53.208777 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:53 crc kubenswrapper[4808]: I1002 16:57:53.399020 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvkks\" (UniqueName: \"kubernetes.io/projected/e976f05e-37b0-4803-bff2-b7b164c0e24d-kube-api-access-nvkks\") pod \"e976f05e-37b0-4803-bff2-b7b164c0e24d\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " Oct 02 16:57:53 crc kubenswrapper[4808]: I1002 16:57:53.399312 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-credential-keys\") pod \"e976f05e-37b0-4803-bff2-b7b164c0e24d\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " Oct 02 16:57:53 crc kubenswrapper[4808]: I1002 16:57:53.399442 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-scripts\") pod \"e976f05e-37b0-4803-bff2-b7b164c0e24d\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " Oct 02 16:57:53 crc kubenswrapper[4808]: I1002 16:57:53.399587 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-fernet-keys\") pod \"e976f05e-37b0-4803-bff2-b7b164c0e24d\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " Oct 02 16:57:53 crc kubenswrapper[4808]: I1002 16:57:53.399785 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-config-data\") pod \"e976f05e-37b0-4803-bff2-b7b164c0e24d\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " Oct 02 16:57:53 crc kubenswrapper[4808]: I1002 16:57:53.399862 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-combined-ca-bundle\") pod \"e976f05e-37b0-4803-bff2-b7b164c0e24d\" (UID: \"e976f05e-37b0-4803-bff2-b7b164c0e24d\") " Oct 02 16:57:53 crc kubenswrapper[4808]: I1002 16:57:53.406005 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-scripts" (OuterVolumeSpecName: "scripts") pod "e976f05e-37b0-4803-bff2-b7b164c0e24d" (UID: "e976f05e-37b0-4803-bff2-b7b164c0e24d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:57:53 crc kubenswrapper[4808]: I1002 16:57:53.406337 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e976f05e-37b0-4803-bff2-b7b164c0e24d-kube-api-access-nvkks" (OuterVolumeSpecName: "kube-api-access-nvkks") pod "e976f05e-37b0-4803-bff2-b7b164c0e24d" (UID: "e976f05e-37b0-4803-bff2-b7b164c0e24d"). InnerVolumeSpecName "kube-api-access-nvkks". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:57:53 crc kubenswrapper[4808]: I1002 16:57:53.406500 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e976f05e-37b0-4803-bff2-b7b164c0e24d" (UID: "e976f05e-37b0-4803-bff2-b7b164c0e24d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:57:53 crc kubenswrapper[4808]: I1002 16:57:53.420506 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "e976f05e-37b0-4803-bff2-b7b164c0e24d" (UID: "e976f05e-37b0-4803-bff2-b7b164c0e24d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:57:53 crc kubenswrapper[4808]: I1002 16:57:53.431936 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-config-data" (OuterVolumeSpecName: "config-data") pod "e976f05e-37b0-4803-bff2-b7b164c0e24d" (UID: "e976f05e-37b0-4803-bff2-b7b164c0e24d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:57:53 crc kubenswrapper[4808]: I1002 16:57:53.432676 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e976f05e-37b0-4803-bff2-b7b164c0e24d" (UID: "e976f05e-37b0-4803-bff2-b7b164c0e24d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:57:53 crc kubenswrapper[4808]: I1002 16:57:53.501997 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:53 crc kubenswrapper[4808]: I1002 16:57:53.502022 4808 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:53 crc kubenswrapper[4808]: I1002 16:57:53.502030 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:53 crc kubenswrapper[4808]: I1002 16:57:53.502040 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:53 crc kubenswrapper[4808]: I1002 16:57:53.502049 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvkks\" (UniqueName: \"kubernetes.io/projected/e976f05e-37b0-4803-bff2-b7b164c0e24d-kube-api-access-nvkks\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:53 crc kubenswrapper[4808]: I1002 16:57:53.502058 4808 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e976f05e-37b0-4803-bff2-b7b164c0e24d-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.007057 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-g8jl2" event={"ID":"e976f05e-37b0-4803-bff2-b7b164c0e24d","Type":"ContainerDied","Data":"500c8bb33f0ef07fd09e291fdcecbd601e1449afb85599056aa8a3c69d8659f2"} Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.007117 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="500c8bb33f0ef07fd09e291fdcecbd601e1449afb85599056aa8a3c69d8659f2" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.007188 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-g8jl2" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.060538 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.128566 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-fgskc"] Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.128833 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" podUID="9aca5301-55c6-4e78-8830-11cb79861ade" containerName="dnsmasq-dns" containerID="cri-o://4418d941a635dd08c35aae363cd8749f5fc4adbad6b78fe5924c550b448290f0" gracePeriod=10 Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.298115 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-g8jl2"] Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.304082 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-g8jl2"] Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.399130 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-ls5jx"] Oct 02 16:57:54 crc kubenswrapper[4808]: E1002 16:57:54.399597 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e976f05e-37b0-4803-bff2-b7b164c0e24d" containerName="keystone-bootstrap" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.399619 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e976f05e-37b0-4803-bff2-b7b164c0e24d" containerName="keystone-bootstrap" Oct 02 16:57:54 crc kubenswrapper[4808]: E1002 16:57:54.399642 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cd07478-020c-4aee-b27f-9a193440b09e" containerName="init" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.399650 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cd07478-020c-4aee-b27f-9a193440b09e" containerName="init" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.399826 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cd07478-020c-4aee-b27f-9a193440b09e" containerName="init" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.399853 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e976f05e-37b0-4803-bff2-b7b164c0e24d" containerName="keystone-bootstrap" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.400544 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.402707 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.402877 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.403082 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-b9ff9" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.403958 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.415952 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ls5jx"] Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.521328 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-config-data\") pod \"keystone-bootstrap-ls5jx\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.521386 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjf6b\" (UniqueName: \"kubernetes.io/projected/60c1cab9-fa93-4572-8eb5-f9682a047645-kube-api-access-bjf6b\") pod \"keystone-bootstrap-ls5jx\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.521476 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-scripts\") pod \"keystone-bootstrap-ls5jx\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.521538 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-fernet-keys\") pod \"keystone-bootstrap-ls5jx\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.521658 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-credential-keys\") pod \"keystone-bootstrap-ls5jx\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.521680 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-combined-ca-bundle\") pod \"keystone-bootstrap-ls5jx\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.623153 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-scripts\") pod \"keystone-bootstrap-ls5jx\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.623217 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-fernet-keys\") pod \"keystone-bootstrap-ls5jx\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.623303 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-credential-keys\") pod \"keystone-bootstrap-ls5jx\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.623329 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-combined-ca-bundle\") pod \"keystone-bootstrap-ls5jx\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.623360 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-config-data\") pod \"keystone-bootstrap-ls5jx\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.623382 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjf6b\" (UniqueName: \"kubernetes.io/projected/60c1cab9-fa93-4572-8eb5-f9682a047645-kube-api-access-bjf6b\") pod \"keystone-bootstrap-ls5jx\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.628961 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-config-data\") pod \"keystone-bootstrap-ls5jx\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.629276 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-combined-ca-bundle\") pod \"keystone-bootstrap-ls5jx\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.629479 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-fernet-keys\") pod \"keystone-bootstrap-ls5jx\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.629593 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-scripts\") pod \"keystone-bootstrap-ls5jx\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.636093 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-credential-keys\") pod \"keystone-bootstrap-ls5jx\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.639382 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjf6b\" (UniqueName: \"kubernetes.io/projected/60c1cab9-fa93-4572-8eb5-f9682a047645-kube-api-access-bjf6b\") pod \"keystone-bootstrap-ls5jx\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:57:54 crc kubenswrapper[4808]: I1002 16:57:54.727816 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:57:55 crc kubenswrapper[4808]: I1002 16:57:55.017100 4808 generic.go:334] "Generic (PLEG): container finished" podID="9aca5301-55c6-4e78-8830-11cb79861ade" containerID="4418d941a635dd08c35aae363cd8749f5fc4adbad6b78fe5924c550b448290f0" exitCode=0 Oct 02 16:57:55 crc kubenswrapper[4808]: I1002 16:57:55.017142 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" event={"ID":"9aca5301-55c6-4e78-8830-11cb79861ade","Type":"ContainerDied","Data":"4418d941a635dd08c35aae363cd8749f5fc4adbad6b78fe5924c550b448290f0"} Oct 02 16:57:55 crc kubenswrapper[4808]: I1002 16:57:55.405888 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e976f05e-37b0-4803-bff2-b7b164c0e24d" path="/var/lib/kubelet/pods/e976f05e-37b0-4803-bff2-b7b164c0e24d/volumes" Oct 02 16:57:58 crc kubenswrapper[4808]: I1002 16:57:58.129830 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" podUID="9aca5301-55c6-4e78-8830-11cb79861ade" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.127:5353: connect: connection refused" Oct 02 16:58:03 crc kubenswrapper[4808]: I1002 16:58:03.128846 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" podUID="9aca5301-55c6-4e78-8830-11cb79861ade" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.127:5353: connect: connection refused" Oct 02 16:58:03 crc kubenswrapper[4808]: E1002 16:58:03.915525 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Oct 02 16:58:03 crc kubenswrapper[4808]: E1002 16:58:03.916217 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jhth4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-ckf4j_openstack(74a1c6ad-0412-4c46-90ce-34869c9ca9e6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 16:58:03 crc kubenswrapper[4808]: E1002 16:58:03.917666 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-ckf4j" podUID="74a1c6ad-0412-4c46-90ce-34869c9ca9e6" Oct 02 16:58:04 crc kubenswrapper[4808]: E1002 16:58:04.093980 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-ckf4j" podUID="74a1c6ad-0412-4c46-90ce-34869c9ca9e6" Oct 02 16:58:05 crc kubenswrapper[4808]: E1002 16:58:05.043536 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 02 16:58:05 crc kubenswrapper[4808]: E1002 16:58:05.044089 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bd29c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-l2fls_openstack(b3d4166a-c265-4cb4-8481-32e6b0144997): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 16:58:05 crc kubenswrapper[4808]: E1002 16:58:05.046040 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-l2fls" podUID="b3d4166a-c265-4cb4-8481-32e6b0144997" Oct 02 16:58:05 crc kubenswrapper[4808]: I1002 16:58:05.135046 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" event={"ID":"9aca5301-55c6-4e78-8830-11cb79861ade","Type":"ContainerDied","Data":"6fa8caf486b55347a70536c195fc44eac7ff259ace4461681392fcd0d2879936"} Oct 02 16:58:05 crc kubenswrapper[4808]: I1002 16:58:05.135476 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fa8caf486b55347a70536c195fc44eac7ff259ace4461681392fcd0d2879936" Oct 02 16:58:05 crc kubenswrapper[4808]: E1002 16:58:05.139206 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-l2fls" podUID="b3d4166a-c265-4cb4-8481-32e6b0144997" Oct 02 16:58:05 crc kubenswrapper[4808]: I1002 16:58:05.140851 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:58:05 crc kubenswrapper[4808]: I1002 16:58:05.229966 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-dns-svc\") pod \"9aca5301-55c6-4e78-8830-11cb79861ade\" (UID: \"9aca5301-55c6-4e78-8830-11cb79861ade\") " Oct 02 16:58:05 crc kubenswrapper[4808]: I1002 16:58:05.230667 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rbj4\" (UniqueName: \"kubernetes.io/projected/9aca5301-55c6-4e78-8830-11cb79861ade-kube-api-access-4rbj4\") pod \"9aca5301-55c6-4e78-8830-11cb79861ade\" (UID: \"9aca5301-55c6-4e78-8830-11cb79861ade\") " Oct 02 16:58:05 crc kubenswrapper[4808]: I1002 16:58:05.230938 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-ovsdbserver-nb\") pod \"9aca5301-55c6-4e78-8830-11cb79861ade\" (UID: \"9aca5301-55c6-4e78-8830-11cb79861ade\") " Oct 02 16:58:05 crc kubenswrapper[4808]: I1002 16:58:05.231083 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-config\") pod \"9aca5301-55c6-4e78-8830-11cb79861ade\" (UID: \"9aca5301-55c6-4e78-8830-11cb79861ade\") " Oct 02 16:58:05 crc kubenswrapper[4808]: I1002 16:58:05.231526 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-ovsdbserver-sb\") pod \"9aca5301-55c6-4e78-8830-11cb79861ade\" (UID: \"9aca5301-55c6-4e78-8830-11cb79861ade\") " Oct 02 16:58:05 crc kubenswrapper[4808]: I1002 16:58:05.249463 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aca5301-55c6-4e78-8830-11cb79861ade-kube-api-access-4rbj4" (OuterVolumeSpecName: "kube-api-access-4rbj4") pod "9aca5301-55c6-4e78-8830-11cb79861ade" (UID: "9aca5301-55c6-4e78-8830-11cb79861ade"). InnerVolumeSpecName "kube-api-access-4rbj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:58:05 crc kubenswrapper[4808]: I1002 16:58:05.269699 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9aca5301-55c6-4e78-8830-11cb79861ade" (UID: "9aca5301-55c6-4e78-8830-11cb79861ade"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:58:05 crc kubenswrapper[4808]: I1002 16:58:05.273902 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9aca5301-55c6-4e78-8830-11cb79861ade" (UID: "9aca5301-55c6-4e78-8830-11cb79861ade"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:58:05 crc kubenswrapper[4808]: I1002 16:58:05.275144 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9aca5301-55c6-4e78-8830-11cb79861ade" (UID: "9aca5301-55c6-4e78-8830-11cb79861ade"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:58:05 crc kubenswrapper[4808]: I1002 16:58:05.280698 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-config" (OuterVolumeSpecName: "config") pod "9aca5301-55c6-4e78-8830-11cb79861ade" (UID: "9aca5301-55c6-4e78-8830-11cb79861ade"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:58:05 crc kubenswrapper[4808]: I1002 16:58:05.333942 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:05 crc kubenswrapper[4808]: I1002 16:58:05.333990 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:05 crc kubenswrapper[4808]: I1002 16:58:05.334007 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:05 crc kubenswrapper[4808]: I1002 16:58:05.334020 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rbj4\" (UniqueName: \"kubernetes.io/projected/9aca5301-55c6-4e78-8830-11cb79861ade-kube-api-access-4rbj4\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:05 crc kubenswrapper[4808]: I1002 16:58:05.334033 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aca5301-55c6-4e78-8830-11cb79861ade-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:05 crc kubenswrapper[4808]: I1002 16:58:05.454469 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ls5jx"] Oct 02 16:58:05 crc kubenswrapper[4808]: W1002 16:58:05.461069 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60c1cab9_fa93_4572_8eb5_f9682a047645.slice/crio-6fa4cd87136f62807351b82c74cc4fb652b857879dccafb6f7018a99a76e92fb WatchSource:0}: Error finding container 6fa4cd87136f62807351b82c74cc4fb652b857879dccafb6f7018a99a76e92fb: Status 404 returned error can't find the container with id 6fa4cd87136f62807351b82c74cc4fb652b857879dccafb6f7018a99a76e92fb Oct 02 16:58:06 crc kubenswrapper[4808]: I1002 16:58:06.147267 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ls5jx" event={"ID":"60c1cab9-fa93-4572-8eb5-f9682a047645","Type":"ContainerStarted","Data":"e5e9e7e471281420db45c65650174a226228535e2ecb32d6ba170bce5321b83c"} Oct 02 16:58:06 crc kubenswrapper[4808]: I1002 16:58:06.147633 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ls5jx" event={"ID":"60c1cab9-fa93-4572-8eb5-f9682a047645","Type":"ContainerStarted","Data":"6fa4cd87136f62807351b82c74cc4fb652b857879dccafb6f7018a99a76e92fb"} Oct 02 16:58:06 crc kubenswrapper[4808]: I1002 16:58:06.154508 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"435bbb18-4b67-4b64-a922-f57717d8adc8","Type":"ContainerStarted","Data":"d7993a7f6691bfbed0900f5a37f318919132778746c9766a0af52e73181c3d1a"} Oct 02 16:58:06 crc kubenswrapper[4808]: I1002 16:58:06.156583 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54f9b7b8d9-fgskc" Oct 02 16:58:06 crc kubenswrapper[4808]: I1002 16:58:06.157546 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dl6kd" event={"ID":"7a437827-755a-4648-a40e-18a8ef89355d","Type":"ContainerStarted","Data":"393a485895bf15b7c626820cf9f585356a32f02586b4781d1e7507f694b63239"} Oct 02 16:58:06 crc kubenswrapper[4808]: I1002 16:58:06.169639 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-ls5jx" podStartSLOduration=12.169623316 podStartE2EDuration="12.169623316s" podCreationTimestamp="2025-10-02 16:57:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:58:06.16313978 +0000 UTC m=+1053.488668780" watchObservedRunningTime="2025-10-02 16:58:06.169623316 +0000 UTC m=+1053.495152316" Oct 02 16:58:06 crc kubenswrapper[4808]: I1002 16:58:06.178502 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-fgskc"] Oct 02 16:58:06 crc kubenswrapper[4808]: I1002 16:58:06.185480 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-fgskc"] Oct 02 16:58:06 crc kubenswrapper[4808]: I1002 16:58:06.192215 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-dl6kd" podStartSLOduration=2.640753194 podStartE2EDuration="23.192192419s" podCreationTimestamp="2025-10-02 16:57:43 +0000 UTC" firstStartedPulling="2025-10-02 16:57:44.440757327 +0000 UTC m=+1031.766286327" lastFinishedPulling="2025-10-02 16:58:04.992196552 +0000 UTC m=+1052.317725552" observedRunningTime="2025-10-02 16:58:06.188887359 +0000 UTC m=+1053.514416439" watchObservedRunningTime="2025-10-02 16:58:06.192192419 +0000 UTC m=+1053.517721429" Oct 02 16:58:07 crc kubenswrapper[4808]: I1002 16:58:07.166199 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"435bbb18-4b67-4b64-a922-f57717d8adc8","Type":"ContainerStarted","Data":"767043d8b360818eb395174c8ee4ab656acd08e5375d55f2ce457b30ec751a9d"} Oct 02 16:58:07 crc kubenswrapper[4808]: I1002 16:58:07.168878 4808 generic.go:334] "Generic (PLEG): container finished" podID="7a437827-755a-4648-a40e-18a8ef89355d" containerID="393a485895bf15b7c626820cf9f585356a32f02586b4781d1e7507f694b63239" exitCode=0 Oct 02 16:58:07 crc kubenswrapper[4808]: I1002 16:58:07.168932 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dl6kd" event={"ID":"7a437827-755a-4648-a40e-18a8ef89355d","Type":"ContainerDied","Data":"393a485895bf15b7c626820cf9f585356a32f02586b4781d1e7507f694b63239"} Oct 02 16:58:07 crc kubenswrapper[4808]: I1002 16:58:07.416546 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aca5301-55c6-4e78-8830-11cb79861ade" path="/var/lib/kubelet/pods/9aca5301-55c6-4e78-8830-11cb79861ade/volumes" Oct 02 16:58:10 crc kubenswrapper[4808]: I1002 16:58:10.212842 4808 generic.go:334] "Generic (PLEG): container finished" podID="60c1cab9-fa93-4572-8eb5-f9682a047645" containerID="e5e9e7e471281420db45c65650174a226228535e2ecb32d6ba170bce5321b83c" exitCode=0 Oct 02 16:58:10 crc kubenswrapper[4808]: I1002 16:58:10.213007 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ls5jx" event={"ID":"60c1cab9-fa93-4572-8eb5-f9682a047645","Type":"ContainerDied","Data":"e5e9e7e471281420db45c65650174a226228535e2ecb32d6ba170bce5321b83c"} Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:11.963911 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dl6kd" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:11.972629 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.086643 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-fernet-keys\") pod \"60c1cab9-fa93-4572-8eb5-f9682a047645\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.086690 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a437827-755a-4648-a40e-18a8ef89355d-scripts\") pod \"7a437827-755a-4648-a40e-18a8ef89355d\" (UID: \"7a437827-755a-4648-a40e-18a8ef89355d\") " Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.086706 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a437827-755a-4648-a40e-18a8ef89355d-config-data\") pod \"7a437827-755a-4648-a40e-18a8ef89355d\" (UID: \"7a437827-755a-4648-a40e-18a8ef89355d\") " Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.086742 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a437827-755a-4648-a40e-18a8ef89355d-logs\") pod \"7a437827-755a-4648-a40e-18a8ef89355d\" (UID: \"7a437827-755a-4648-a40e-18a8ef89355d\") " Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.086814 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a437827-755a-4648-a40e-18a8ef89355d-combined-ca-bundle\") pod \"7a437827-755a-4648-a40e-18a8ef89355d\" (UID: \"7a437827-755a-4648-a40e-18a8ef89355d\") " Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.086840 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfntz\" (UniqueName: \"kubernetes.io/projected/7a437827-755a-4648-a40e-18a8ef89355d-kube-api-access-wfntz\") pod \"7a437827-755a-4648-a40e-18a8ef89355d\" (UID: \"7a437827-755a-4648-a40e-18a8ef89355d\") " Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.086880 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-combined-ca-bundle\") pod \"60c1cab9-fa93-4572-8eb5-f9682a047645\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.086959 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-config-data\") pod \"60c1cab9-fa93-4572-8eb5-f9682a047645\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.086987 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjf6b\" (UniqueName: \"kubernetes.io/projected/60c1cab9-fa93-4572-8eb5-f9682a047645-kube-api-access-bjf6b\") pod \"60c1cab9-fa93-4572-8eb5-f9682a047645\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.087001 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-scripts\") pod \"60c1cab9-fa93-4572-8eb5-f9682a047645\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.087019 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-credential-keys\") pod \"60c1cab9-fa93-4572-8eb5-f9682a047645\" (UID: \"60c1cab9-fa93-4572-8eb5-f9682a047645\") " Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.088782 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a437827-755a-4648-a40e-18a8ef89355d-logs" (OuterVolumeSpecName: "logs") pod "7a437827-755a-4648-a40e-18a8ef89355d" (UID: "7a437827-755a-4648-a40e-18a8ef89355d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.093750 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "60c1cab9-fa93-4572-8eb5-f9682a047645" (UID: "60c1cab9-fa93-4572-8eb5-f9682a047645"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.093750 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a437827-755a-4648-a40e-18a8ef89355d-scripts" (OuterVolumeSpecName: "scripts") pod "7a437827-755a-4648-a40e-18a8ef89355d" (UID: "7a437827-755a-4648-a40e-18a8ef89355d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.093842 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-scripts" (OuterVolumeSpecName: "scripts") pod "60c1cab9-fa93-4572-8eb5-f9682a047645" (UID: "60c1cab9-fa93-4572-8eb5-f9682a047645"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.101406 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "60c1cab9-fa93-4572-8eb5-f9682a047645" (UID: "60c1cab9-fa93-4572-8eb5-f9682a047645"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.101433 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60c1cab9-fa93-4572-8eb5-f9682a047645-kube-api-access-bjf6b" (OuterVolumeSpecName: "kube-api-access-bjf6b") pod "60c1cab9-fa93-4572-8eb5-f9682a047645" (UID: "60c1cab9-fa93-4572-8eb5-f9682a047645"). InnerVolumeSpecName "kube-api-access-bjf6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.101477 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a437827-755a-4648-a40e-18a8ef89355d-kube-api-access-wfntz" (OuterVolumeSpecName: "kube-api-access-wfntz") pod "7a437827-755a-4648-a40e-18a8ef89355d" (UID: "7a437827-755a-4648-a40e-18a8ef89355d"). InnerVolumeSpecName "kube-api-access-wfntz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.115501 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a437827-755a-4648-a40e-18a8ef89355d-config-data" (OuterVolumeSpecName: "config-data") pod "7a437827-755a-4648-a40e-18a8ef89355d" (UID: "7a437827-755a-4648-a40e-18a8ef89355d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.116989 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60c1cab9-fa93-4572-8eb5-f9682a047645" (UID: "60c1cab9-fa93-4572-8eb5-f9682a047645"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.117081 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-config-data" (OuterVolumeSpecName: "config-data") pod "60c1cab9-fa93-4572-8eb5-f9682a047645" (UID: "60c1cab9-fa93-4572-8eb5-f9682a047645"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.117400 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a437827-755a-4648-a40e-18a8ef89355d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a437827-755a-4648-a40e-18a8ef89355d" (UID: "7a437827-755a-4648-a40e-18a8ef89355d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.189277 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.189318 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjf6b\" (UniqueName: \"kubernetes.io/projected/60c1cab9-fa93-4572-8eb5-f9682a047645-kube-api-access-bjf6b\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.189332 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.189341 4808 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.189349 4808 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.189358 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a437827-755a-4648-a40e-18a8ef89355d-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.189365 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a437827-755a-4648-a40e-18a8ef89355d-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.189375 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a437827-755a-4648-a40e-18a8ef89355d-logs\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.189384 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a437827-755a-4648-a40e-18a8ef89355d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.189395 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfntz\" (UniqueName: \"kubernetes.io/projected/7a437827-755a-4648-a40e-18a8ef89355d-kube-api-access-wfntz\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.189403 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60c1cab9-fa93-4572-8eb5-f9682a047645-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.231039 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ls5jx" event={"ID":"60c1cab9-fa93-4572-8eb5-f9682a047645","Type":"ContainerDied","Data":"6fa4cd87136f62807351b82c74cc4fb652b857879dccafb6f7018a99a76e92fb"} Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.231072 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fa4cd87136f62807351b82c74cc4fb652b857879dccafb6f7018a99a76e92fb" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.231125 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ls5jx" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.236591 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"435bbb18-4b67-4b64-a922-f57717d8adc8","Type":"ContainerStarted","Data":"c77562ca22229ccb03ca743a8776eae3231f405a55857f55f534cc676e0acf37"} Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.238889 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dl6kd" event={"ID":"7a437827-755a-4648-a40e-18a8ef89355d","Type":"ContainerDied","Data":"fc7c521aba60bd23b39c3b2578eae342f5f48e3148fabfc107e086fe1104a1ad"} Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.238914 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc7c521aba60bd23b39c3b2578eae342f5f48e3148fabfc107e086fe1104a1ad" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.238972 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dl6kd" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.316247 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-577d5bdf64-z4hwp"] Oct 02 16:58:12 crc kubenswrapper[4808]: E1002 16:58:12.316790 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aca5301-55c6-4e78-8830-11cb79861ade" containerName="init" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.316808 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aca5301-55c6-4e78-8830-11cb79861ade" containerName="init" Oct 02 16:58:12 crc kubenswrapper[4808]: E1002 16:58:12.316824 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aca5301-55c6-4e78-8830-11cb79861ade" containerName="dnsmasq-dns" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.316830 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aca5301-55c6-4e78-8830-11cb79861ade" containerName="dnsmasq-dns" Oct 02 16:58:12 crc kubenswrapper[4808]: E1002 16:58:12.316842 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60c1cab9-fa93-4572-8eb5-f9682a047645" containerName="keystone-bootstrap" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.316850 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="60c1cab9-fa93-4572-8eb5-f9682a047645" containerName="keystone-bootstrap" Oct 02 16:58:12 crc kubenswrapper[4808]: E1002 16:58:12.316883 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a437827-755a-4648-a40e-18a8ef89355d" containerName="placement-db-sync" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.316891 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a437827-755a-4648-a40e-18a8ef89355d" containerName="placement-db-sync" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.317077 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aca5301-55c6-4e78-8830-11cb79861ade" containerName="dnsmasq-dns" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.317180 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a437827-755a-4648-a40e-18a8ef89355d" containerName="placement-db-sync" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.317195 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="60c1cab9-fa93-4572-8eb5-f9682a047645" containerName="keystone-bootstrap" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.317840 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.335445 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.336009 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.336731 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.337355 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.337490 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.337555 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-b9ff9" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.356041 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-577d5bdf64-z4hwp"] Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.495435 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-scripts\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.495485 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-config-data\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.495531 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-fernet-keys\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.495555 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-combined-ca-bundle\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.495576 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgrtl\" (UniqueName: \"kubernetes.io/projected/9300a6c4-e3c6-4cef-98bf-192b9d5549be-kube-api-access-wgrtl\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.495773 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-public-tls-certs\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.495985 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-internal-tls-certs\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.496052 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-credential-keys\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.597495 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-scripts\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.597550 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-config-data\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.597578 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-fernet-keys\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.597606 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-combined-ca-bundle\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.597630 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgrtl\" (UniqueName: \"kubernetes.io/projected/9300a6c4-e3c6-4cef-98bf-192b9d5549be-kube-api-access-wgrtl\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.597657 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-public-tls-certs\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.597715 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-internal-tls-certs\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.597743 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-credential-keys\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.603693 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-internal-tls-certs\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.604118 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-scripts\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.605756 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-public-tls-certs\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.606106 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-credential-keys\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.606981 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-config-data\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.607327 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-combined-ca-bundle\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.607554 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9300a6c4-e3c6-4cef-98bf-192b9d5549be-fernet-keys\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.613444 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgrtl\" (UniqueName: \"kubernetes.io/projected/9300a6c4-e3c6-4cef-98bf-192b9d5549be-kube-api-access-wgrtl\") pod \"keystone-577d5bdf64-z4hwp\" (UID: \"9300a6c4-e3c6-4cef-98bf-192b9d5549be\") " pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:12 crc kubenswrapper[4808]: I1002 16:58:12.653619 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.079827 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7dbcfb7656-7hqb7"] Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.081247 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.083261 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.083818 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-wbbc4" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.084011 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.085782 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.087350 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.103558 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7dbcfb7656-7hqb7"] Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.139631 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-577d5bdf64-z4hwp"] Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.207555 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6317ce08-fda5-4271-908c-a18a0fd29180-internal-tls-certs\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.212838 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6317ce08-fda5-4271-908c-a18a0fd29180-logs\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.212920 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6317ce08-fda5-4271-908c-a18a0fd29180-public-tls-certs\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.213268 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6317ce08-fda5-4271-908c-a18a0fd29180-combined-ca-bundle\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.213319 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6317ce08-fda5-4271-908c-a18a0fd29180-scripts\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.213552 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg284\" (UniqueName: \"kubernetes.io/projected/6317ce08-fda5-4271-908c-a18a0fd29180-kube-api-access-gg284\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.213609 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6317ce08-fda5-4271-908c-a18a0fd29180-config-data\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.248814 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-577d5bdf64-z4hwp" event={"ID":"9300a6c4-e3c6-4cef-98bf-192b9d5549be","Type":"ContainerStarted","Data":"976a9524d4edf4e41ee03c0b3d9682c2c72ec8b8c99c1c97cda5c52e34f9f836"} Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.314659 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6317ce08-fda5-4271-908c-a18a0fd29180-combined-ca-bundle\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.314700 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6317ce08-fda5-4271-908c-a18a0fd29180-scripts\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.314763 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg284\" (UniqueName: \"kubernetes.io/projected/6317ce08-fda5-4271-908c-a18a0fd29180-kube-api-access-gg284\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.314784 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6317ce08-fda5-4271-908c-a18a0fd29180-config-data\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.314809 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6317ce08-fda5-4271-908c-a18a0fd29180-internal-tls-certs\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.314854 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6317ce08-fda5-4271-908c-a18a0fd29180-logs\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.314880 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6317ce08-fda5-4271-908c-a18a0fd29180-public-tls-certs\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.315925 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6317ce08-fda5-4271-908c-a18a0fd29180-logs\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.318866 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6317ce08-fda5-4271-908c-a18a0fd29180-scripts\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.319791 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6317ce08-fda5-4271-908c-a18a0fd29180-internal-tls-certs\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.320505 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6317ce08-fda5-4271-908c-a18a0fd29180-combined-ca-bundle\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.322108 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6317ce08-fda5-4271-908c-a18a0fd29180-config-data\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.327406 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6317ce08-fda5-4271-908c-a18a0fd29180-public-tls-certs\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.331491 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg284\" (UniqueName: \"kubernetes.io/projected/6317ce08-fda5-4271-908c-a18a0fd29180-kube-api-access-gg284\") pod \"placement-7dbcfb7656-7hqb7\" (UID: \"6317ce08-fda5-4271-908c-a18a0fd29180\") " pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.398559 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:13 crc kubenswrapper[4808]: I1002 16:58:13.661454 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7dbcfb7656-7hqb7"] Oct 02 16:58:13 crc kubenswrapper[4808]: W1002 16:58:13.677069 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6317ce08_fda5_4271_908c_a18a0fd29180.slice/crio-b59c6a05b7a414b9bb11e470b40f07b9573f6bb95dde2a5166ea4cbea24232e0 WatchSource:0}: Error finding container b59c6a05b7a414b9bb11e470b40f07b9573f6bb95dde2a5166ea4cbea24232e0: Status 404 returned error can't find the container with id b59c6a05b7a414b9bb11e470b40f07b9573f6bb95dde2a5166ea4cbea24232e0 Oct 02 16:58:14 crc kubenswrapper[4808]: I1002 16:58:14.260324 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7dbcfb7656-7hqb7" event={"ID":"6317ce08-fda5-4271-908c-a18a0fd29180","Type":"ContainerStarted","Data":"f78cff7ad51d5cd628ad10b0b4d5a4e63a33598f32d3dfbb4a28c2e6b6ef922e"} Oct 02 16:58:14 crc kubenswrapper[4808]: I1002 16:58:14.260577 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7dbcfb7656-7hqb7" event={"ID":"6317ce08-fda5-4271-908c-a18a0fd29180","Type":"ContainerStarted","Data":"052e2013f96b20ed921005efc675163d237ac80fc2bb41bcf39796f3d55d5294"} Oct 02 16:58:14 crc kubenswrapper[4808]: I1002 16:58:14.260588 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7dbcfb7656-7hqb7" event={"ID":"6317ce08-fda5-4271-908c-a18a0fd29180","Type":"ContainerStarted","Data":"b59c6a05b7a414b9bb11e470b40f07b9573f6bb95dde2a5166ea4cbea24232e0"} Oct 02 16:58:14 crc kubenswrapper[4808]: I1002 16:58:14.260716 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:14 crc kubenswrapper[4808]: I1002 16:58:14.263588 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-577d5bdf64-z4hwp" event={"ID":"9300a6c4-e3c6-4cef-98bf-192b9d5549be","Type":"ContainerStarted","Data":"40e59b869ea0255d848c35ebed7b3d802e1fbad4377c21a81ef8ac53cf324d7d"} Oct 02 16:58:14 crc kubenswrapper[4808]: I1002 16:58:14.264214 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:14 crc kubenswrapper[4808]: I1002 16:58:14.265754 4808 generic.go:334] "Generic (PLEG): container finished" podID="c50129b6-292e-434b-a617-2781dac4101e" containerID="e11d95d6c344e8547933c8bf98688be3e6c5e1d7a02e78850fed4d4acf0bf0ed" exitCode=0 Oct 02 16:58:14 crc kubenswrapper[4808]: I1002 16:58:14.265776 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-zpkmw" event={"ID":"c50129b6-292e-434b-a617-2781dac4101e","Type":"ContainerDied","Data":"e11d95d6c344e8547933c8bf98688be3e6c5e1d7a02e78850fed4d4acf0bf0ed"} Oct 02 16:58:14 crc kubenswrapper[4808]: I1002 16:58:14.282508 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7dbcfb7656-7hqb7" podStartSLOduration=1.282486104 podStartE2EDuration="1.282486104s" podCreationTimestamp="2025-10-02 16:58:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:58:14.281946869 +0000 UTC m=+1061.607475869" watchObservedRunningTime="2025-10-02 16:58:14.282486104 +0000 UTC m=+1061.608015104" Oct 02 16:58:14 crc kubenswrapper[4808]: I1002 16:58:14.334317 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-577d5bdf64-z4hwp" podStartSLOduration=2.33429946 podStartE2EDuration="2.33429946s" podCreationTimestamp="2025-10-02 16:58:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:58:14.332682826 +0000 UTC m=+1061.658211846" watchObservedRunningTime="2025-10-02 16:58:14.33429946 +0000 UTC m=+1061.659828460" Oct 02 16:58:15 crc kubenswrapper[4808]: I1002 16:58:15.273682 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:18 crc kubenswrapper[4808]: I1002 16:58:18.740321 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 16:58:18 crc kubenswrapper[4808]: I1002 16:58:18.740850 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 16:58:18 crc kubenswrapper[4808]: I1002 16:58:18.740901 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 16:58:18 crc kubenswrapper[4808]: I1002 16:58:18.741687 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2d9c8f9e3115f67c6ef88d1a8e4688cf6f5cba6faca4d6c650065b45bf9b7b1b"} pod="openshift-machine-config-operator/machine-config-daemon-7z66r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 16:58:18 crc kubenswrapper[4808]: I1002 16:58:18.741750 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" containerID="cri-o://2d9c8f9e3115f67c6ef88d1a8e4688cf6f5cba6faca4d6c650065b45bf9b7b1b" gracePeriod=600 Oct 02 16:58:18 crc kubenswrapper[4808]: I1002 16:58:18.947516 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-zpkmw" Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.121537 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c50129b6-292e-434b-a617-2781dac4101e-combined-ca-bundle\") pod \"c50129b6-292e-434b-a617-2781dac4101e\" (UID: \"c50129b6-292e-434b-a617-2781dac4101e\") " Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.121604 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcqrq\" (UniqueName: \"kubernetes.io/projected/c50129b6-292e-434b-a617-2781dac4101e-kube-api-access-qcqrq\") pod \"c50129b6-292e-434b-a617-2781dac4101e\" (UID: \"c50129b6-292e-434b-a617-2781dac4101e\") " Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.121658 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c50129b6-292e-434b-a617-2781dac4101e-config\") pod \"c50129b6-292e-434b-a617-2781dac4101e\" (UID: \"c50129b6-292e-434b-a617-2781dac4101e\") " Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.128434 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c50129b6-292e-434b-a617-2781dac4101e-kube-api-access-qcqrq" (OuterVolumeSpecName: "kube-api-access-qcqrq") pod "c50129b6-292e-434b-a617-2781dac4101e" (UID: "c50129b6-292e-434b-a617-2781dac4101e"). InnerVolumeSpecName "kube-api-access-qcqrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.176681 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c50129b6-292e-434b-a617-2781dac4101e-config" (OuterVolumeSpecName: "config") pod "c50129b6-292e-434b-a617-2781dac4101e" (UID: "c50129b6-292e-434b-a617-2781dac4101e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.177640 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c50129b6-292e-434b-a617-2781dac4101e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c50129b6-292e-434b-a617-2781dac4101e" (UID: "c50129b6-292e-434b-a617-2781dac4101e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.223273 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c50129b6-292e-434b-a617-2781dac4101e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.223308 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcqrq\" (UniqueName: \"kubernetes.io/projected/c50129b6-292e-434b-a617-2781dac4101e-kube-api-access-qcqrq\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.223322 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c50129b6-292e-434b-a617-2781dac4101e-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.314778 4808 generic.go:334] "Generic (PLEG): container finished" podID="f83ce425-101d-4489-94a4-5c256eb29328" containerID="2d9c8f9e3115f67c6ef88d1a8e4688cf6f5cba6faca4d6c650065b45bf9b7b1b" exitCode=0 Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.314887 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" event={"ID":"f83ce425-101d-4489-94a4-5c256eb29328","Type":"ContainerDied","Data":"2d9c8f9e3115f67c6ef88d1a8e4688cf6f5cba6faca4d6c650065b45bf9b7b1b"} Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.314953 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" event={"ID":"f83ce425-101d-4489-94a4-5c256eb29328","Type":"ContainerStarted","Data":"325982695f20979226cf6b1b8bfe2ce4f159293f5eb31e61d627777c496fdb01"} Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.314973 4808 scope.go:117] "RemoveContainer" containerID="8d253701a5ddc087ee8a36b225b0284b63257d65500bbf63a65e988cba6e1fa1" Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.326026 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"435bbb18-4b67-4b64-a922-f57717d8adc8","Type":"ContainerStarted","Data":"0957eee368d34a70abc6a0a2af7daab10f521f3dc14d9f7f524e960b7f4fe496"} Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.326183 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="435bbb18-4b67-4b64-a922-f57717d8adc8" containerName="ceilometer-central-agent" containerID="cri-o://d7993a7f6691bfbed0900f5a37f318919132778746c9766a0af52e73181c3d1a" gracePeriod=30 Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.328166 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.331398 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="435bbb18-4b67-4b64-a922-f57717d8adc8" containerName="proxy-httpd" containerID="cri-o://0957eee368d34a70abc6a0a2af7daab10f521f3dc14d9f7f524e960b7f4fe496" gracePeriod=30 Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.331610 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="435bbb18-4b67-4b64-a922-f57717d8adc8" containerName="sg-core" containerID="cri-o://c77562ca22229ccb03ca743a8776eae3231f405a55857f55f534cc676e0acf37" gracePeriod=30 Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.331655 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="435bbb18-4b67-4b64-a922-f57717d8adc8" containerName="ceilometer-notification-agent" containerID="cri-o://767043d8b360818eb395174c8ee4ab656acd08e5375d55f2ce457b30ec751a9d" gracePeriod=30 Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.344448 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-zpkmw" event={"ID":"c50129b6-292e-434b-a617-2781dac4101e","Type":"ContainerDied","Data":"0634d96b8fdf947f50283e7c7b55a7becfb92b15117cbc3c3c6d7ec534ece3e4"} Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.344524 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0634d96b8fdf947f50283e7c7b55a7becfb92b15117cbc3c3c6d7ec534ece3e4" Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.345567 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-zpkmw" Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.354954 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ckf4j" event={"ID":"74a1c6ad-0412-4c46-90ce-34869c9ca9e6","Type":"ContainerStarted","Data":"3004ddbbb72e926e110a2e3f850d513f465f8b92dd7eb9e67b1ac75f7d9d4580"} Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.356773 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9762094239999999 podStartE2EDuration="36.356754635s" podCreationTimestamp="2025-10-02 16:57:43 +0000 UTC" firstStartedPulling="2025-10-02 16:57:44.597307198 +0000 UTC m=+1031.922836198" lastFinishedPulling="2025-10-02 16:58:18.977852409 +0000 UTC m=+1066.303381409" observedRunningTime="2025-10-02 16:58:19.349758195 +0000 UTC m=+1066.675287195" watchObservedRunningTime="2025-10-02 16:58:19.356754635 +0000 UTC m=+1066.682283635" Oct 02 16:58:19 crc kubenswrapper[4808]: I1002 16:58:19.378221 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-ckf4j" podStartSLOduration=2.020090154 podStartE2EDuration="36.378200057s" podCreationTimestamp="2025-10-02 16:57:43 +0000 UTC" firstStartedPulling="2025-10-02 16:57:44.611972615 +0000 UTC m=+1031.937501615" lastFinishedPulling="2025-10-02 16:58:18.970082508 +0000 UTC m=+1066.295611518" observedRunningTime="2025-10-02 16:58:19.376065459 +0000 UTC m=+1066.701594449" watchObservedRunningTime="2025-10-02 16:58:19.378200057 +0000 UTC m=+1066.703729067" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.223293 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-d44pp"] Oct 02 16:58:20 crc kubenswrapper[4808]: E1002 16:58:20.223984 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c50129b6-292e-434b-a617-2781dac4101e" containerName="neutron-db-sync" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.223997 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c50129b6-292e-434b-a617-2781dac4101e" containerName="neutron-db-sync" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.224157 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c50129b6-292e-434b-a617-2781dac4101e" containerName="neutron-db-sync" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.228720 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b946d459c-d44pp" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.249008 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-d44pp"] Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.290283 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-59d95756b-rl9bd"] Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.291938 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.299606 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.299770 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.299951 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-cprk7" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.300175 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.314953 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-59d95756b-rl9bd"] Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.342520 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-ovsdbserver-sb\") pod \"dnsmasq-dns-7b946d459c-d44pp\" (UID: \"ad76984a-1396-4144-8f61-664cd77cb5aa\") " pod="openstack/dnsmasq-dns-7b946d459c-d44pp" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.342584 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-ovsdbserver-nb\") pod \"dnsmasq-dns-7b946d459c-d44pp\" (UID: \"ad76984a-1396-4144-8f61-664cd77cb5aa\") " pod="openstack/dnsmasq-dns-7b946d459c-d44pp" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.342610 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-dns-svc\") pod \"dnsmasq-dns-7b946d459c-d44pp\" (UID: \"ad76984a-1396-4144-8f61-664cd77cb5aa\") " pod="openstack/dnsmasq-dns-7b946d459c-d44pp" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.342627 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hgf7\" (UniqueName: \"kubernetes.io/projected/ad76984a-1396-4144-8f61-664cd77cb5aa-kube-api-access-8hgf7\") pod \"dnsmasq-dns-7b946d459c-d44pp\" (UID: \"ad76984a-1396-4144-8f61-664cd77cb5aa\") " pod="openstack/dnsmasq-dns-7b946d459c-d44pp" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.342899 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-config\") pod \"dnsmasq-dns-7b946d459c-d44pp\" (UID: \"ad76984a-1396-4144-8f61-664cd77cb5aa\") " pod="openstack/dnsmasq-dns-7b946d459c-d44pp" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.366044 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-l2fls" event={"ID":"b3d4166a-c265-4cb4-8481-32e6b0144997","Type":"ContainerStarted","Data":"e5ffb3b4bc1236be9b7bac9a4a7c6a0bdd5c0340bfc95e1c5eeed99afeae8907"} Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.368648 4808 generic.go:334] "Generic (PLEG): container finished" podID="435bbb18-4b67-4b64-a922-f57717d8adc8" containerID="0957eee368d34a70abc6a0a2af7daab10f521f3dc14d9f7f524e960b7f4fe496" exitCode=0 Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.368670 4808 generic.go:334] "Generic (PLEG): container finished" podID="435bbb18-4b67-4b64-a922-f57717d8adc8" containerID="c77562ca22229ccb03ca743a8776eae3231f405a55857f55f534cc676e0acf37" exitCode=2 Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.368677 4808 generic.go:334] "Generic (PLEG): container finished" podID="435bbb18-4b67-4b64-a922-f57717d8adc8" containerID="d7993a7f6691bfbed0900f5a37f318919132778746c9766a0af52e73181c3d1a" exitCode=0 Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.368718 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"435bbb18-4b67-4b64-a922-f57717d8adc8","Type":"ContainerDied","Data":"0957eee368d34a70abc6a0a2af7daab10f521f3dc14d9f7f524e960b7f4fe496"} Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.368736 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"435bbb18-4b67-4b64-a922-f57717d8adc8","Type":"ContainerDied","Data":"c77562ca22229ccb03ca743a8776eae3231f405a55857f55f534cc676e0acf37"} Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.368745 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"435bbb18-4b67-4b64-a922-f57717d8adc8","Type":"ContainerDied","Data":"d7993a7f6691bfbed0900f5a37f318919132778746c9766a0af52e73181c3d1a"} Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.384690 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-l2fls" podStartSLOduration=2.654960782 podStartE2EDuration="37.38467216s" podCreationTimestamp="2025-10-02 16:57:43 +0000 UTC" firstStartedPulling="2025-10-02 16:57:44.222429423 +0000 UTC m=+1031.547958423" lastFinishedPulling="2025-10-02 16:58:18.952140801 +0000 UTC m=+1066.277669801" observedRunningTime="2025-10-02 16:58:20.381055072 +0000 UTC m=+1067.706584092" watchObservedRunningTime="2025-10-02 16:58:20.38467216 +0000 UTC m=+1067.710201160" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.444534 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-ovsdbserver-sb\") pod \"dnsmasq-dns-7b946d459c-d44pp\" (UID: \"ad76984a-1396-4144-8f61-664cd77cb5aa\") " pod="openstack/dnsmasq-dns-7b946d459c-d44pp" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.444618 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-ovsdbserver-nb\") pod \"dnsmasq-dns-7b946d459c-d44pp\" (UID: \"ad76984a-1396-4144-8f61-664cd77cb5aa\") " pod="openstack/dnsmasq-dns-7b946d459c-d44pp" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.444652 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-dns-svc\") pod \"dnsmasq-dns-7b946d459c-d44pp\" (UID: \"ad76984a-1396-4144-8f61-664cd77cb5aa\") " pod="openstack/dnsmasq-dns-7b946d459c-d44pp" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.444676 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-combined-ca-bundle\") pod \"neutron-59d95756b-rl9bd\" (UID: \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\") " pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.444713 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hgf7\" (UniqueName: \"kubernetes.io/projected/ad76984a-1396-4144-8f61-664cd77cb5aa-kube-api-access-8hgf7\") pod \"dnsmasq-dns-7b946d459c-d44pp\" (UID: \"ad76984a-1396-4144-8f61-664cd77cb5aa\") " pod="openstack/dnsmasq-dns-7b946d459c-d44pp" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.444773 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-config\") pod \"dnsmasq-dns-7b946d459c-d44pp\" (UID: \"ad76984a-1396-4144-8f61-664cd77cb5aa\") " pod="openstack/dnsmasq-dns-7b946d459c-d44pp" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.444813 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-config\") pod \"neutron-59d95756b-rl9bd\" (UID: \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\") " pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.444858 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqw2p\" (UniqueName: \"kubernetes.io/projected/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-kube-api-access-vqw2p\") pod \"neutron-59d95756b-rl9bd\" (UID: \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\") " pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.444883 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-httpd-config\") pod \"neutron-59d95756b-rl9bd\" (UID: \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\") " pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.444984 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-ovndb-tls-certs\") pod \"neutron-59d95756b-rl9bd\" (UID: \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\") " pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.445828 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-dns-svc\") pod \"dnsmasq-dns-7b946d459c-d44pp\" (UID: \"ad76984a-1396-4144-8f61-664cd77cb5aa\") " pod="openstack/dnsmasq-dns-7b946d459c-d44pp" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.446337 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-ovsdbserver-sb\") pod \"dnsmasq-dns-7b946d459c-d44pp\" (UID: \"ad76984a-1396-4144-8f61-664cd77cb5aa\") " pod="openstack/dnsmasq-dns-7b946d459c-d44pp" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.446396 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-config\") pod \"dnsmasq-dns-7b946d459c-d44pp\" (UID: \"ad76984a-1396-4144-8f61-664cd77cb5aa\") " pod="openstack/dnsmasq-dns-7b946d459c-d44pp" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.446999 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-ovsdbserver-nb\") pod \"dnsmasq-dns-7b946d459c-d44pp\" (UID: \"ad76984a-1396-4144-8f61-664cd77cb5aa\") " pod="openstack/dnsmasq-dns-7b946d459c-d44pp" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.467519 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hgf7\" (UniqueName: \"kubernetes.io/projected/ad76984a-1396-4144-8f61-664cd77cb5aa-kube-api-access-8hgf7\") pod \"dnsmasq-dns-7b946d459c-d44pp\" (UID: \"ad76984a-1396-4144-8f61-664cd77cb5aa\") " pod="openstack/dnsmasq-dns-7b946d459c-d44pp" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.546495 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b946d459c-d44pp" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.546954 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-ovndb-tls-certs\") pod \"neutron-59d95756b-rl9bd\" (UID: \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\") " pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.547040 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-combined-ca-bundle\") pod \"neutron-59d95756b-rl9bd\" (UID: \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\") " pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.547144 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-config\") pod \"neutron-59d95756b-rl9bd\" (UID: \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\") " pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.547181 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqw2p\" (UniqueName: \"kubernetes.io/projected/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-kube-api-access-vqw2p\") pod \"neutron-59d95756b-rl9bd\" (UID: \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\") " pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.547204 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-httpd-config\") pod \"neutron-59d95756b-rl9bd\" (UID: \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\") " pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.550581 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-combined-ca-bundle\") pod \"neutron-59d95756b-rl9bd\" (UID: \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\") " pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.551103 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-ovndb-tls-certs\") pod \"neutron-59d95756b-rl9bd\" (UID: \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\") " pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.551516 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-config\") pod \"neutron-59d95756b-rl9bd\" (UID: \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\") " pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.569177 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqw2p\" (UniqueName: \"kubernetes.io/projected/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-kube-api-access-vqw2p\") pod \"neutron-59d95756b-rl9bd\" (UID: \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\") " pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.582669 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-httpd-config\") pod \"neutron-59d95756b-rl9bd\" (UID: \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\") " pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:20 crc kubenswrapper[4808]: I1002 16:58:20.632779 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:21 crc kubenswrapper[4808]: I1002 16:58:21.093715 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-d44pp"] Oct 02 16:58:21 crc kubenswrapper[4808]: W1002 16:58:21.102833 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad76984a_1396_4144_8f61_664cd77cb5aa.slice/crio-ca2b81ce2b5124d39e9933b428a4478504582263914a112d991421af411042bb WatchSource:0}: Error finding container ca2b81ce2b5124d39e9933b428a4478504582263914a112d991421af411042bb: Status 404 returned error can't find the container with id ca2b81ce2b5124d39e9933b428a4478504582263914a112d991421af411042bb Oct 02 16:58:21 crc kubenswrapper[4808]: I1002 16:58:21.380723 4808 generic.go:334] "Generic (PLEG): container finished" podID="ad76984a-1396-4144-8f61-664cd77cb5aa" containerID="e4ef75e980f4fffe5411e2b4df718abdfb17ca2b9a085c71bd5fbf1ded03e95a" exitCode=0 Oct 02 16:58:21 crc kubenswrapper[4808]: I1002 16:58:21.380796 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-d44pp" event={"ID":"ad76984a-1396-4144-8f61-664cd77cb5aa","Type":"ContainerDied","Data":"e4ef75e980f4fffe5411e2b4df718abdfb17ca2b9a085c71bd5fbf1ded03e95a"} Oct 02 16:58:21 crc kubenswrapper[4808]: I1002 16:58:21.381055 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-d44pp" event={"ID":"ad76984a-1396-4144-8f61-664cd77cb5aa","Type":"ContainerStarted","Data":"ca2b81ce2b5124d39e9933b428a4478504582263914a112d991421af411042bb"} Oct 02 16:58:21 crc kubenswrapper[4808]: I1002 16:58:21.561829 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-59d95756b-rl9bd"] Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.393078 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-d44pp" event={"ID":"ad76984a-1396-4144-8f61-664cd77cb5aa","Type":"ContainerStarted","Data":"2c31739baa1672ba3aba3c1eab11c077378d920ea415d3964edf6753b64df227"} Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.395713 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b946d459c-d44pp" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.397361 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59d95756b-rl9bd" event={"ID":"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d","Type":"ContainerStarted","Data":"7da26f1db3010f3f7f174cddb5acd13403301a6470274ecef4756758ffdc9612"} Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.397414 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59d95756b-rl9bd" event={"ID":"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d","Type":"ContainerStarted","Data":"1472f9b375176e20444d88bf3d34fe527360069ebe368cd9ad0f391466df3b5b"} Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.397430 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59d95756b-rl9bd" event={"ID":"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d","Type":"ContainerStarted","Data":"3874ed3d128e268bb5d1dd244afd11d6de52e6b8aeb76e8fbcb15ce335f81297"} Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.398099 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.399539 4808 generic.go:334] "Generic (PLEG): container finished" podID="74a1c6ad-0412-4c46-90ce-34869c9ca9e6" containerID="3004ddbbb72e926e110a2e3f850d513f465f8b92dd7eb9e67b1ac75f7d9d4580" exitCode=0 Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.399618 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ckf4j" event={"ID":"74a1c6ad-0412-4c46-90ce-34869c9ca9e6","Type":"ContainerDied","Data":"3004ddbbb72e926e110a2e3f850d513f465f8b92dd7eb9e67b1ac75f7d9d4580"} Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.416006 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b946d459c-d44pp" podStartSLOduration=2.415986203 podStartE2EDuration="2.415986203s" podCreationTimestamp="2025-10-02 16:58:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:58:22.412008485 +0000 UTC m=+1069.737537495" watchObservedRunningTime="2025-10-02 16:58:22.415986203 +0000 UTC m=+1069.741515203" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.458175 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-59d95756b-rl9bd" podStartSLOduration=2.458156988 podStartE2EDuration="2.458156988s" podCreationTimestamp="2025-10-02 16:58:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:58:22.445739821 +0000 UTC m=+1069.771268811" watchObservedRunningTime="2025-10-02 16:58:22.458156988 +0000 UTC m=+1069.783685988" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.493261 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-78b67777b5-h245x"] Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.494940 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.498503 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.504025 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-78b67777b5-h245x"] Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.515215 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.681146 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c31691-57ad-4fa3-9fad-b81d7655d443-internal-tls-certs\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.681207 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/72c31691-57ad-4fa3-9fad-b81d7655d443-config\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.681280 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c31691-57ad-4fa3-9fad-b81d7655d443-combined-ca-bundle\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.681305 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c31691-57ad-4fa3-9fad-b81d7655d443-ovndb-tls-certs\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.681377 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcd8c\" (UniqueName: \"kubernetes.io/projected/72c31691-57ad-4fa3-9fad-b81d7655d443-kube-api-access-bcd8c\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.681422 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c31691-57ad-4fa3-9fad-b81d7655d443-public-tls-certs\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.681438 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/72c31691-57ad-4fa3-9fad-b81d7655d443-httpd-config\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.782617 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c31691-57ad-4fa3-9fad-b81d7655d443-internal-tls-certs\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.782682 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/72c31691-57ad-4fa3-9fad-b81d7655d443-config\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.782731 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c31691-57ad-4fa3-9fad-b81d7655d443-combined-ca-bundle\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.782750 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c31691-57ad-4fa3-9fad-b81d7655d443-ovndb-tls-certs\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.782815 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcd8c\" (UniqueName: \"kubernetes.io/projected/72c31691-57ad-4fa3-9fad-b81d7655d443-kube-api-access-bcd8c\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.782861 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c31691-57ad-4fa3-9fad-b81d7655d443-public-tls-certs\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.782885 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/72c31691-57ad-4fa3-9fad-b81d7655d443-httpd-config\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.788499 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c31691-57ad-4fa3-9fad-b81d7655d443-combined-ca-bundle\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.788847 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c31691-57ad-4fa3-9fad-b81d7655d443-ovndb-tls-certs\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.789018 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c31691-57ad-4fa3-9fad-b81d7655d443-public-tls-certs\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.800719 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcd8c\" (UniqueName: \"kubernetes.io/projected/72c31691-57ad-4fa3-9fad-b81d7655d443-kube-api-access-bcd8c\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.802935 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/72c31691-57ad-4fa3-9fad-b81d7655d443-httpd-config\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.804273 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c31691-57ad-4fa3-9fad-b81d7655d443-internal-tls-certs\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.809173 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/72c31691-57ad-4fa3-9fad-b81d7655d443-config\") pod \"neutron-78b67777b5-h245x\" (UID: \"72c31691-57ad-4fa3-9fad-b81d7655d443\") " pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:22 crc kubenswrapper[4808]: I1002 16:58:22.824689 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:23 crc kubenswrapper[4808]: I1002 16:58:23.521157 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-78b67777b5-h245x"] Oct 02 16:58:23 crc kubenswrapper[4808]: I1002 16:58:23.688499 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ckf4j" Oct 02 16:58:23 crc kubenswrapper[4808]: I1002 16:58:23.800932 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a1c6ad-0412-4c46-90ce-34869c9ca9e6-combined-ca-bundle\") pod \"74a1c6ad-0412-4c46-90ce-34869c9ca9e6\" (UID: \"74a1c6ad-0412-4c46-90ce-34869c9ca9e6\") " Oct 02 16:58:23 crc kubenswrapper[4808]: I1002 16:58:23.801284 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/74a1c6ad-0412-4c46-90ce-34869c9ca9e6-db-sync-config-data\") pod \"74a1c6ad-0412-4c46-90ce-34869c9ca9e6\" (UID: \"74a1c6ad-0412-4c46-90ce-34869c9ca9e6\") " Oct 02 16:58:23 crc kubenswrapper[4808]: I1002 16:58:23.801323 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhth4\" (UniqueName: \"kubernetes.io/projected/74a1c6ad-0412-4c46-90ce-34869c9ca9e6-kube-api-access-jhth4\") pod \"74a1c6ad-0412-4c46-90ce-34869c9ca9e6\" (UID: \"74a1c6ad-0412-4c46-90ce-34869c9ca9e6\") " Oct 02 16:58:23 crc kubenswrapper[4808]: I1002 16:58:23.804976 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74a1c6ad-0412-4c46-90ce-34869c9ca9e6-kube-api-access-jhth4" (OuterVolumeSpecName: "kube-api-access-jhth4") pod "74a1c6ad-0412-4c46-90ce-34869c9ca9e6" (UID: "74a1c6ad-0412-4c46-90ce-34869c9ca9e6"). InnerVolumeSpecName "kube-api-access-jhth4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:58:23 crc kubenswrapper[4808]: I1002 16:58:23.805799 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74a1c6ad-0412-4c46-90ce-34869c9ca9e6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "74a1c6ad-0412-4c46-90ce-34869c9ca9e6" (UID: "74a1c6ad-0412-4c46-90ce-34869c9ca9e6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:23 crc kubenswrapper[4808]: I1002 16:58:23.825769 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74a1c6ad-0412-4c46-90ce-34869c9ca9e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74a1c6ad-0412-4c46-90ce-34869c9ca9e6" (UID: "74a1c6ad-0412-4c46-90ce-34869c9ca9e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:23 crc kubenswrapper[4808]: I1002 16:58:23.903578 4808 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/74a1c6ad-0412-4c46-90ce-34869c9ca9e6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:23 crc kubenswrapper[4808]: I1002 16:58:23.903608 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhth4\" (UniqueName: \"kubernetes.io/projected/74a1c6ad-0412-4c46-90ce-34869c9ca9e6-kube-api-access-jhth4\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:23 crc kubenswrapper[4808]: I1002 16:58:23.903620 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a1c6ad-0412-4c46-90ce-34869c9ca9e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:23 crc kubenswrapper[4808]: I1002 16:58:23.969777 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.106115 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-combined-ca-bundle\") pod \"435bbb18-4b67-4b64-a922-f57717d8adc8\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.106188 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-config-data\") pod \"435bbb18-4b67-4b64-a922-f57717d8adc8\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.106226 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-scripts\") pod \"435bbb18-4b67-4b64-a922-f57717d8adc8\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.106278 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtfq8\" (UniqueName: \"kubernetes.io/projected/435bbb18-4b67-4b64-a922-f57717d8adc8-kube-api-access-xtfq8\") pod \"435bbb18-4b67-4b64-a922-f57717d8adc8\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.106347 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-sg-core-conf-yaml\") pod \"435bbb18-4b67-4b64-a922-f57717d8adc8\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.106409 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/435bbb18-4b67-4b64-a922-f57717d8adc8-log-httpd\") pod \"435bbb18-4b67-4b64-a922-f57717d8adc8\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.106435 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/435bbb18-4b67-4b64-a922-f57717d8adc8-run-httpd\") pod \"435bbb18-4b67-4b64-a922-f57717d8adc8\" (UID: \"435bbb18-4b67-4b64-a922-f57717d8adc8\") " Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.107045 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/435bbb18-4b67-4b64-a922-f57717d8adc8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "435bbb18-4b67-4b64-a922-f57717d8adc8" (UID: "435bbb18-4b67-4b64-a922-f57717d8adc8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.133598 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/435bbb18-4b67-4b64-a922-f57717d8adc8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "435bbb18-4b67-4b64-a922-f57717d8adc8" (UID: "435bbb18-4b67-4b64-a922-f57717d8adc8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.133752 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-scripts" (OuterVolumeSpecName: "scripts") pod "435bbb18-4b67-4b64-a922-f57717d8adc8" (UID: "435bbb18-4b67-4b64-a922-f57717d8adc8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.135386 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/435bbb18-4b67-4b64-a922-f57717d8adc8-kube-api-access-xtfq8" (OuterVolumeSpecName: "kube-api-access-xtfq8") pod "435bbb18-4b67-4b64-a922-f57717d8adc8" (UID: "435bbb18-4b67-4b64-a922-f57717d8adc8"). InnerVolumeSpecName "kube-api-access-xtfq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.213015 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "435bbb18-4b67-4b64-a922-f57717d8adc8" (UID: "435bbb18-4b67-4b64-a922-f57717d8adc8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.215739 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.215763 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtfq8\" (UniqueName: \"kubernetes.io/projected/435bbb18-4b67-4b64-a922-f57717d8adc8-kube-api-access-xtfq8\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.215772 4808 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.215779 4808 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/435bbb18-4b67-4b64-a922-f57717d8adc8-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.215787 4808 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/435bbb18-4b67-4b64-a922-f57717d8adc8-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.247179 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "435bbb18-4b67-4b64-a922-f57717d8adc8" (UID: "435bbb18-4b67-4b64-a922-f57717d8adc8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.263511 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-config-data" (OuterVolumeSpecName: "config-data") pod "435bbb18-4b67-4b64-a922-f57717d8adc8" (UID: "435bbb18-4b67-4b64-a922-f57717d8adc8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.317225 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.317279 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/435bbb18-4b67-4b64-a922-f57717d8adc8-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.422080 4808 generic.go:334] "Generic (PLEG): container finished" podID="b3d4166a-c265-4cb4-8481-32e6b0144997" containerID="e5ffb3b4bc1236be9b7bac9a4a7c6a0bdd5c0340bfc95e1c5eeed99afeae8907" exitCode=0 Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.422171 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-l2fls" event={"ID":"b3d4166a-c265-4cb4-8481-32e6b0144997","Type":"ContainerDied","Data":"e5ffb3b4bc1236be9b7bac9a4a7c6a0bdd5c0340bfc95e1c5eeed99afeae8907"} Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.424931 4808 generic.go:334] "Generic (PLEG): container finished" podID="435bbb18-4b67-4b64-a922-f57717d8adc8" containerID="767043d8b360818eb395174c8ee4ab656acd08e5375d55f2ce457b30ec751a9d" exitCode=0 Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.424985 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.424990 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"435bbb18-4b67-4b64-a922-f57717d8adc8","Type":"ContainerDied","Data":"767043d8b360818eb395174c8ee4ab656acd08e5375d55f2ce457b30ec751a9d"} Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.425015 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"435bbb18-4b67-4b64-a922-f57717d8adc8","Type":"ContainerDied","Data":"57a9233bdfbfe05a8a59069e6e6442792909c1c901539f75f70adf637b12f350"} Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.425032 4808 scope.go:117] "RemoveContainer" containerID="0957eee368d34a70abc6a0a2af7daab10f521f3dc14d9f7f524e960b7f4fe496" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.427576 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78b67777b5-h245x" event={"ID":"72c31691-57ad-4fa3-9fad-b81d7655d443","Type":"ContainerStarted","Data":"cfe5121002c1948d109d175a186a9304b81e09e9149fcb5d2a922efcc1c174c8"} Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.427601 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78b67777b5-h245x" event={"ID":"72c31691-57ad-4fa3-9fad-b81d7655d443","Type":"ContainerStarted","Data":"bc0a748fad30df3c39654ae4201e2ddd824d44cf88f3da0faded2f7ea1ca9219"} Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.427611 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78b67777b5-h245x" event={"ID":"72c31691-57ad-4fa3-9fad-b81d7655d443","Type":"ContainerStarted","Data":"ffd9630d99372f04c4ba7e241099c4614a96866517983013f3d777f0c27de21b"} Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.427790 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.429395 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ckf4j" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.429416 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ckf4j" event={"ID":"74a1c6ad-0412-4c46-90ce-34869c9ca9e6","Type":"ContainerDied","Data":"9ad521c143efbef049450fa0e7f2ea88874eb7a29c46eeb91bbeca7ca65cb823"} Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.429450 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ad521c143efbef049450fa0e7f2ea88874eb7a29c46eeb91bbeca7ca65cb823" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.462387 4808 scope.go:117] "RemoveContainer" containerID="c77562ca22229ccb03ca743a8776eae3231f405a55857f55f534cc676e0acf37" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.482211 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-78b67777b5-h245x" podStartSLOduration=2.482194675 podStartE2EDuration="2.482194675s" podCreationTimestamp="2025-10-02 16:58:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:58:24.467811434 +0000 UTC m=+1071.793340424" watchObservedRunningTime="2025-10-02 16:58:24.482194675 +0000 UTC m=+1071.807723675" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.491819 4808 scope.go:117] "RemoveContainer" containerID="767043d8b360818eb395174c8ee4ab656acd08e5375d55f2ce457b30ec751a9d" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.497030 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.508910 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.516453 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:58:24 crc kubenswrapper[4808]: E1002 16:58:24.516866 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="435bbb18-4b67-4b64-a922-f57717d8adc8" containerName="proxy-httpd" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.516881 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="435bbb18-4b67-4b64-a922-f57717d8adc8" containerName="proxy-httpd" Oct 02 16:58:24 crc kubenswrapper[4808]: E1002 16:58:24.516899 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="435bbb18-4b67-4b64-a922-f57717d8adc8" containerName="ceilometer-notification-agent" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.516904 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="435bbb18-4b67-4b64-a922-f57717d8adc8" containerName="ceilometer-notification-agent" Oct 02 16:58:24 crc kubenswrapper[4808]: E1002 16:58:24.516920 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="435bbb18-4b67-4b64-a922-f57717d8adc8" containerName="sg-core" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.516926 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="435bbb18-4b67-4b64-a922-f57717d8adc8" containerName="sg-core" Oct 02 16:58:24 crc kubenswrapper[4808]: E1002 16:58:24.516936 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a1c6ad-0412-4c46-90ce-34869c9ca9e6" containerName="barbican-db-sync" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.516942 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a1c6ad-0412-4c46-90ce-34869c9ca9e6" containerName="barbican-db-sync" Oct 02 16:58:24 crc kubenswrapper[4808]: E1002 16:58:24.516950 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="435bbb18-4b67-4b64-a922-f57717d8adc8" containerName="ceilometer-central-agent" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.516955 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="435bbb18-4b67-4b64-a922-f57717d8adc8" containerName="ceilometer-central-agent" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.517092 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="435bbb18-4b67-4b64-a922-f57717d8adc8" containerName="sg-core" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.517113 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="74a1c6ad-0412-4c46-90ce-34869c9ca9e6" containerName="barbican-db-sync" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.517128 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="435bbb18-4b67-4b64-a922-f57717d8adc8" containerName="proxy-httpd" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.517150 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="435bbb18-4b67-4b64-a922-f57717d8adc8" containerName="ceilometer-notification-agent" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.517163 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="435bbb18-4b67-4b64-a922-f57717d8adc8" containerName="ceilometer-central-agent" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.519445 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.523977 4808 scope.go:117] "RemoveContainer" containerID="d7993a7f6691bfbed0900f5a37f318919132778746c9766a0af52e73181c3d1a" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.525801 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.526174 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.585369 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.606469 4808 scope.go:117] "RemoveContainer" containerID="0957eee368d34a70abc6a0a2af7daab10f521f3dc14d9f7f524e960b7f4fe496" Oct 02 16:58:24 crc kubenswrapper[4808]: E1002 16:58:24.608589 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0957eee368d34a70abc6a0a2af7daab10f521f3dc14d9f7f524e960b7f4fe496\": container with ID starting with 0957eee368d34a70abc6a0a2af7daab10f521f3dc14d9f7f524e960b7f4fe496 not found: ID does not exist" containerID="0957eee368d34a70abc6a0a2af7daab10f521f3dc14d9f7f524e960b7f4fe496" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.608865 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0957eee368d34a70abc6a0a2af7daab10f521f3dc14d9f7f524e960b7f4fe496"} err="failed to get container status \"0957eee368d34a70abc6a0a2af7daab10f521f3dc14d9f7f524e960b7f4fe496\": rpc error: code = NotFound desc = could not find container \"0957eee368d34a70abc6a0a2af7daab10f521f3dc14d9f7f524e960b7f4fe496\": container with ID starting with 0957eee368d34a70abc6a0a2af7daab10f521f3dc14d9f7f524e960b7f4fe496 not found: ID does not exist" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.608895 4808 scope.go:117] "RemoveContainer" containerID="c77562ca22229ccb03ca743a8776eae3231f405a55857f55f534cc676e0acf37" Oct 02 16:58:24 crc kubenswrapper[4808]: E1002 16:58:24.610051 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c77562ca22229ccb03ca743a8776eae3231f405a55857f55f534cc676e0acf37\": container with ID starting with c77562ca22229ccb03ca743a8776eae3231f405a55857f55f534cc676e0acf37 not found: ID does not exist" containerID="c77562ca22229ccb03ca743a8776eae3231f405a55857f55f534cc676e0acf37" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.611139 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c77562ca22229ccb03ca743a8776eae3231f405a55857f55f534cc676e0acf37"} err="failed to get container status \"c77562ca22229ccb03ca743a8776eae3231f405a55857f55f534cc676e0acf37\": rpc error: code = NotFound desc = could not find container \"c77562ca22229ccb03ca743a8776eae3231f405a55857f55f534cc676e0acf37\": container with ID starting with c77562ca22229ccb03ca743a8776eae3231f405a55857f55f534cc676e0acf37 not found: ID does not exist" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.611343 4808 scope.go:117] "RemoveContainer" containerID="767043d8b360818eb395174c8ee4ab656acd08e5375d55f2ce457b30ec751a9d" Oct 02 16:58:24 crc kubenswrapper[4808]: E1002 16:58:24.622008 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"767043d8b360818eb395174c8ee4ab656acd08e5375d55f2ce457b30ec751a9d\": container with ID starting with 767043d8b360818eb395174c8ee4ab656acd08e5375d55f2ce457b30ec751a9d not found: ID does not exist" containerID="767043d8b360818eb395174c8ee4ab656acd08e5375d55f2ce457b30ec751a9d" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.622051 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"767043d8b360818eb395174c8ee4ab656acd08e5375d55f2ce457b30ec751a9d"} err="failed to get container status \"767043d8b360818eb395174c8ee4ab656acd08e5375d55f2ce457b30ec751a9d\": rpc error: code = NotFound desc = could not find container \"767043d8b360818eb395174c8ee4ab656acd08e5375d55f2ce457b30ec751a9d\": container with ID starting with 767043d8b360818eb395174c8ee4ab656acd08e5375d55f2ce457b30ec751a9d not found: ID does not exist" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.622075 4808 scope.go:117] "RemoveContainer" containerID="d7993a7f6691bfbed0900f5a37f318919132778746c9766a0af52e73181c3d1a" Oct 02 16:58:24 crc kubenswrapper[4808]: E1002 16:58:24.624194 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7993a7f6691bfbed0900f5a37f318919132778746c9766a0af52e73181c3d1a\": container with ID starting with d7993a7f6691bfbed0900f5a37f318919132778746c9766a0af52e73181c3d1a not found: ID does not exist" containerID="d7993a7f6691bfbed0900f5a37f318919132778746c9766a0af52e73181c3d1a" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.624248 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7993a7f6691bfbed0900f5a37f318919132778746c9766a0af52e73181c3d1a"} err="failed to get container status \"d7993a7f6691bfbed0900f5a37f318919132778746c9766a0af52e73181c3d1a\": rpc error: code = NotFound desc = could not find container \"d7993a7f6691bfbed0900f5a37f318919132778746c9766a0af52e73181c3d1a\": container with ID starting with d7993a7f6691bfbed0900f5a37f318919132778746c9766a0af52e73181c3d1a not found: ID does not exist" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.625382 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c16fcba0-e224-4b8f-a220-71184caeede4-run-httpd\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.625473 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp4w2\" (UniqueName: \"kubernetes.io/projected/c16fcba0-e224-4b8f-a220-71184caeede4-kube-api-access-dp4w2\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.625517 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.625532 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.625594 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c16fcba0-e224-4b8f-a220-71184caeede4-log-httpd\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.625753 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-config-data\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.625818 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-scripts\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.644274 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5b4f6b689c-zbtm2"] Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.659913 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5b4f6b689c-zbtm2" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.665163 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-54854ffb46-kktlh"] Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.665269 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.665465 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.665564 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-mfvng" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.666458 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.670744 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.686303 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-54854ffb46-kktlh"] Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.711295 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5b4f6b689c-zbtm2"] Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.729153 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c16fcba0-e224-4b8f-a220-71184caeede4-run-httpd\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.729203 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp4w2\" (UniqueName: \"kubernetes.io/projected/c16fcba0-e224-4b8f-a220-71184caeede4-kube-api-access-dp4w2\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.729223 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.729253 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.729280 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c16fcba0-e224-4b8f-a220-71184caeede4-log-httpd\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.729356 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-config-data\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.729386 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-scripts\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.730063 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c16fcba0-e224-4b8f-a220-71184caeede4-run-httpd\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.732587 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c16fcba0-e224-4b8f-a220-71184caeede4-log-httpd\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.734435 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-scripts\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.735475 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.736622 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.736680 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-d44pp"] Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.736888 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-config-data\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.736884 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b946d459c-d44pp" podUID="ad76984a-1396-4144-8f61-664cd77cb5aa" containerName="dnsmasq-dns" containerID="cri-o://2c31739baa1672ba3aba3c1eab11c077378d920ea415d3964edf6753b64df227" gracePeriod=10 Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.751400 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp4w2\" (UniqueName: \"kubernetes.io/projected/c16fcba0-e224-4b8f-a220-71184caeede4-kube-api-access-dp4w2\") pod \"ceilometer-0\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.760306 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-gb8s6"] Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.761873 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.772840 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-gb8s6"] Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.806290 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-798c55655d-t754s"] Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.807746 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.816139 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-798c55655d-t754s"] Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.816512 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.831158 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/695150b9-231f-4f3b-8e53-47b067a55025-logs\") pod \"barbican-worker-5b4f6b689c-zbtm2\" (UID: \"695150b9-231f-4f3b-8e53-47b067a55025\") " pod="openstack/barbican-worker-5b4f6b689c-zbtm2" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.831199 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccq8k\" (UniqueName: \"kubernetes.io/projected/f36f9202-6bbb-404f-865d-206a87585936-kube-api-access-ccq8k\") pod \"barbican-keystone-listener-54854ffb46-kktlh\" (UID: \"f36f9202-6bbb-404f-865d-206a87585936\") " pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.831250 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f36f9202-6bbb-404f-865d-206a87585936-config-data-custom\") pod \"barbican-keystone-listener-54854ffb46-kktlh\" (UID: \"f36f9202-6bbb-404f-865d-206a87585936\") " pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.831300 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f36f9202-6bbb-404f-865d-206a87585936-combined-ca-bundle\") pod \"barbican-keystone-listener-54854ffb46-kktlh\" (UID: \"f36f9202-6bbb-404f-865d-206a87585936\") " pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.831332 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/695150b9-231f-4f3b-8e53-47b067a55025-config-data-custom\") pod \"barbican-worker-5b4f6b689c-zbtm2\" (UID: \"695150b9-231f-4f3b-8e53-47b067a55025\") " pod="openstack/barbican-worker-5b4f6b689c-zbtm2" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.831379 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/695150b9-231f-4f3b-8e53-47b067a55025-combined-ca-bundle\") pod \"barbican-worker-5b4f6b689c-zbtm2\" (UID: \"695150b9-231f-4f3b-8e53-47b067a55025\") " pod="openstack/barbican-worker-5b4f6b689c-zbtm2" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.831407 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhkcx\" (UniqueName: \"kubernetes.io/projected/695150b9-231f-4f3b-8e53-47b067a55025-kube-api-access-vhkcx\") pod \"barbican-worker-5b4f6b689c-zbtm2\" (UID: \"695150b9-231f-4f3b-8e53-47b067a55025\") " pod="openstack/barbican-worker-5b4f6b689c-zbtm2" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.831426 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f36f9202-6bbb-404f-865d-206a87585936-logs\") pod \"barbican-keystone-listener-54854ffb46-kktlh\" (UID: \"f36f9202-6bbb-404f-865d-206a87585936\") " pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.831439 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/695150b9-231f-4f3b-8e53-47b067a55025-config-data\") pod \"barbican-worker-5b4f6b689c-zbtm2\" (UID: \"695150b9-231f-4f3b-8e53-47b067a55025\") " pod="openstack/barbican-worker-5b4f6b689c-zbtm2" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.831484 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f36f9202-6bbb-404f-865d-206a87585936-config-data\") pod \"barbican-keystone-listener-54854ffb46-kktlh\" (UID: \"f36f9202-6bbb-404f-865d-206a87585936\") " pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.880399 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.933352 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/695150b9-231f-4f3b-8e53-47b067a55025-logs\") pod \"barbican-worker-5b4f6b689c-zbtm2\" (UID: \"695150b9-231f-4f3b-8e53-47b067a55025\") " pod="openstack/barbican-worker-5b4f6b689c-zbtm2" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.933388 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccq8k\" (UniqueName: \"kubernetes.io/projected/f36f9202-6bbb-404f-865d-206a87585936-kube-api-access-ccq8k\") pod \"barbican-keystone-listener-54854ffb46-kktlh\" (UID: \"f36f9202-6bbb-404f-865d-206a87585936\") " pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.933416 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsc5h\" (UniqueName: \"kubernetes.io/projected/e4ed4b3e-896b-4d66-b340-deecd1501807-kube-api-access-gsc5h\") pod \"dnsmasq-dns-6bb684768f-gb8s6\" (UID: \"e4ed4b3e-896b-4d66-b340-deecd1501807\") " pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.933432 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28edd780-f606-40e4-a9a3-babce79c3f0f-logs\") pod \"barbican-api-798c55655d-t754s\" (UID: \"28edd780-f606-40e4-a9a3-babce79c3f0f\") " pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.933454 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb684768f-gb8s6\" (UID: \"e4ed4b3e-896b-4d66-b340-deecd1501807\") " pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.933480 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f36f9202-6bbb-404f-865d-206a87585936-config-data-custom\") pod \"barbican-keystone-listener-54854ffb46-kktlh\" (UID: \"f36f9202-6bbb-404f-865d-206a87585936\") " pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.933499 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28edd780-f606-40e4-a9a3-babce79c3f0f-combined-ca-bundle\") pod \"barbican-api-798c55655d-t754s\" (UID: \"28edd780-f606-40e4-a9a3-babce79c3f0f\") " pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.933515 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mf6v\" (UniqueName: \"kubernetes.io/projected/28edd780-f606-40e4-a9a3-babce79c3f0f-kube-api-access-8mf6v\") pod \"barbican-api-798c55655d-t754s\" (UID: \"28edd780-f606-40e4-a9a3-babce79c3f0f\") " pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.933530 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f36f9202-6bbb-404f-865d-206a87585936-combined-ca-bundle\") pod \"barbican-keystone-listener-54854ffb46-kktlh\" (UID: \"f36f9202-6bbb-404f-865d-206a87585936\") " pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.933560 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/695150b9-231f-4f3b-8e53-47b067a55025-config-data-custom\") pod \"barbican-worker-5b4f6b689c-zbtm2\" (UID: \"695150b9-231f-4f3b-8e53-47b067a55025\") " pod="openstack/barbican-worker-5b4f6b689c-zbtm2" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.933581 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28edd780-f606-40e4-a9a3-babce79c3f0f-config-data\") pod \"barbican-api-798c55655d-t754s\" (UID: \"28edd780-f606-40e4-a9a3-babce79c3f0f\") " pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.933599 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-config\") pod \"dnsmasq-dns-6bb684768f-gb8s6\" (UID: \"e4ed4b3e-896b-4d66-b340-deecd1501807\") " pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.933627 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/695150b9-231f-4f3b-8e53-47b067a55025-combined-ca-bundle\") pod \"barbican-worker-5b4f6b689c-zbtm2\" (UID: \"695150b9-231f-4f3b-8e53-47b067a55025\") " pod="openstack/barbican-worker-5b4f6b689c-zbtm2" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.933645 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-dns-svc\") pod \"dnsmasq-dns-6bb684768f-gb8s6\" (UID: \"e4ed4b3e-896b-4d66-b340-deecd1501807\") " pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.933664 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb684768f-gb8s6\" (UID: \"e4ed4b3e-896b-4d66-b340-deecd1501807\") " pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.933688 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhkcx\" (UniqueName: \"kubernetes.io/projected/695150b9-231f-4f3b-8e53-47b067a55025-kube-api-access-vhkcx\") pod \"barbican-worker-5b4f6b689c-zbtm2\" (UID: \"695150b9-231f-4f3b-8e53-47b067a55025\") " pod="openstack/barbican-worker-5b4f6b689c-zbtm2" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.933705 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/695150b9-231f-4f3b-8e53-47b067a55025-config-data\") pod \"barbican-worker-5b4f6b689c-zbtm2\" (UID: \"695150b9-231f-4f3b-8e53-47b067a55025\") " pod="openstack/barbican-worker-5b4f6b689c-zbtm2" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.933719 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f36f9202-6bbb-404f-865d-206a87585936-logs\") pod \"barbican-keystone-listener-54854ffb46-kktlh\" (UID: \"f36f9202-6bbb-404f-865d-206a87585936\") " pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.933728 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/695150b9-231f-4f3b-8e53-47b067a55025-logs\") pod \"barbican-worker-5b4f6b689c-zbtm2\" (UID: \"695150b9-231f-4f3b-8e53-47b067a55025\") " pod="openstack/barbican-worker-5b4f6b689c-zbtm2" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.933753 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/28edd780-f606-40e4-a9a3-babce79c3f0f-config-data-custom\") pod \"barbican-api-798c55655d-t754s\" (UID: \"28edd780-f606-40e4-a9a3-babce79c3f0f\") " pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.934336 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f36f9202-6bbb-404f-865d-206a87585936-logs\") pod \"barbican-keystone-listener-54854ffb46-kktlh\" (UID: \"f36f9202-6bbb-404f-865d-206a87585936\") " pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.935756 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f36f9202-6bbb-404f-865d-206a87585936-config-data\") pod \"barbican-keystone-listener-54854ffb46-kktlh\" (UID: \"f36f9202-6bbb-404f-865d-206a87585936\") " pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.938263 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f36f9202-6bbb-404f-865d-206a87585936-combined-ca-bundle\") pod \"barbican-keystone-listener-54854ffb46-kktlh\" (UID: \"f36f9202-6bbb-404f-865d-206a87585936\") " pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.939285 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f36f9202-6bbb-404f-865d-206a87585936-config-data-custom\") pod \"barbican-keystone-listener-54854ffb46-kktlh\" (UID: \"f36f9202-6bbb-404f-865d-206a87585936\") " pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.939291 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/695150b9-231f-4f3b-8e53-47b067a55025-config-data\") pod \"barbican-worker-5b4f6b689c-zbtm2\" (UID: \"695150b9-231f-4f3b-8e53-47b067a55025\") " pod="openstack/barbican-worker-5b4f6b689c-zbtm2" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.939598 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/695150b9-231f-4f3b-8e53-47b067a55025-config-data-custom\") pod \"barbican-worker-5b4f6b689c-zbtm2\" (UID: \"695150b9-231f-4f3b-8e53-47b067a55025\") " pod="openstack/barbican-worker-5b4f6b689c-zbtm2" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.940348 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f36f9202-6bbb-404f-865d-206a87585936-config-data\") pod \"barbican-keystone-listener-54854ffb46-kktlh\" (UID: \"f36f9202-6bbb-404f-865d-206a87585936\") " pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.941359 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/695150b9-231f-4f3b-8e53-47b067a55025-combined-ca-bundle\") pod \"barbican-worker-5b4f6b689c-zbtm2\" (UID: \"695150b9-231f-4f3b-8e53-47b067a55025\") " pod="openstack/barbican-worker-5b4f6b689c-zbtm2" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.951488 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhkcx\" (UniqueName: \"kubernetes.io/projected/695150b9-231f-4f3b-8e53-47b067a55025-kube-api-access-vhkcx\") pod \"barbican-worker-5b4f6b689c-zbtm2\" (UID: \"695150b9-231f-4f3b-8e53-47b067a55025\") " pod="openstack/barbican-worker-5b4f6b689c-zbtm2" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.952977 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccq8k\" (UniqueName: \"kubernetes.io/projected/f36f9202-6bbb-404f-865d-206a87585936-kube-api-access-ccq8k\") pod \"barbican-keystone-listener-54854ffb46-kktlh\" (UID: \"f36f9202-6bbb-404f-865d-206a87585936\") " pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" Oct 02 16:58:24 crc kubenswrapper[4808]: I1002 16:58:24.987894 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5b4f6b689c-zbtm2" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.008685 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.039419 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28edd780-f606-40e4-a9a3-babce79c3f0f-config-data\") pod \"barbican-api-798c55655d-t754s\" (UID: \"28edd780-f606-40e4-a9a3-babce79c3f0f\") " pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.039459 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-config\") pod \"dnsmasq-dns-6bb684768f-gb8s6\" (UID: \"e4ed4b3e-896b-4d66-b340-deecd1501807\") " pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.039494 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-dns-svc\") pod \"dnsmasq-dns-6bb684768f-gb8s6\" (UID: \"e4ed4b3e-896b-4d66-b340-deecd1501807\") " pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.039512 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb684768f-gb8s6\" (UID: \"e4ed4b3e-896b-4d66-b340-deecd1501807\") " pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.039564 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/28edd780-f606-40e4-a9a3-babce79c3f0f-config-data-custom\") pod \"barbican-api-798c55655d-t754s\" (UID: \"28edd780-f606-40e4-a9a3-babce79c3f0f\") " pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.039607 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsc5h\" (UniqueName: \"kubernetes.io/projected/e4ed4b3e-896b-4d66-b340-deecd1501807-kube-api-access-gsc5h\") pod \"dnsmasq-dns-6bb684768f-gb8s6\" (UID: \"e4ed4b3e-896b-4d66-b340-deecd1501807\") " pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.039623 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28edd780-f606-40e4-a9a3-babce79c3f0f-logs\") pod \"barbican-api-798c55655d-t754s\" (UID: \"28edd780-f606-40e4-a9a3-babce79c3f0f\") " pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.039644 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb684768f-gb8s6\" (UID: \"e4ed4b3e-896b-4d66-b340-deecd1501807\") " pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.039674 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28edd780-f606-40e4-a9a3-babce79c3f0f-combined-ca-bundle\") pod \"barbican-api-798c55655d-t754s\" (UID: \"28edd780-f606-40e4-a9a3-babce79c3f0f\") " pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.039688 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mf6v\" (UniqueName: \"kubernetes.io/projected/28edd780-f606-40e4-a9a3-babce79c3f0f-kube-api-access-8mf6v\") pod \"barbican-api-798c55655d-t754s\" (UID: \"28edd780-f606-40e4-a9a3-babce79c3f0f\") " pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.040392 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28edd780-f606-40e4-a9a3-babce79c3f0f-logs\") pod \"barbican-api-798c55655d-t754s\" (UID: \"28edd780-f606-40e4-a9a3-babce79c3f0f\") " pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.040853 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb684768f-gb8s6\" (UID: \"e4ed4b3e-896b-4d66-b340-deecd1501807\") " pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.041271 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-config\") pod \"dnsmasq-dns-6bb684768f-gb8s6\" (UID: \"e4ed4b3e-896b-4d66-b340-deecd1501807\") " pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.041606 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb684768f-gb8s6\" (UID: \"e4ed4b3e-896b-4d66-b340-deecd1501807\") " pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.042164 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-dns-svc\") pod \"dnsmasq-dns-6bb684768f-gb8s6\" (UID: \"e4ed4b3e-896b-4d66-b340-deecd1501807\") " pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.044003 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/28edd780-f606-40e4-a9a3-babce79c3f0f-config-data-custom\") pod \"barbican-api-798c55655d-t754s\" (UID: \"28edd780-f606-40e4-a9a3-babce79c3f0f\") " pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.046790 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28edd780-f606-40e4-a9a3-babce79c3f0f-combined-ca-bundle\") pod \"barbican-api-798c55655d-t754s\" (UID: \"28edd780-f606-40e4-a9a3-babce79c3f0f\") " pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.047649 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28edd780-f606-40e4-a9a3-babce79c3f0f-config-data\") pod \"barbican-api-798c55655d-t754s\" (UID: \"28edd780-f606-40e4-a9a3-babce79c3f0f\") " pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.056704 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mf6v\" (UniqueName: \"kubernetes.io/projected/28edd780-f606-40e4-a9a3-babce79c3f0f-kube-api-access-8mf6v\") pod \"barbican-api-798c55655d-t754s\" (UID: \"28edd780-f606-40e4-a9a3-babce79c3f0f\") " pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.061816 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsc5h\" (UniqueName: \"kubernetes.io/projected/e4ed4b3e-896b-4d66-b340-deecd1501807-kube-api-access-gsc5h\") pod \"dnsmasq-dns-6bb684768f-gb8s6\" (UID: \"e4ed4b3e-896b-4d66-b340-deecd1501807\") " pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.065534 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.088738 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.315973 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b946d459c-d44pp" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.375578 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hgf7\" (UniqueName: \"kubernetes.io/projected/ad76984a-1396-4144-8f61-664cd77cb5aa-kube-api-access-8hgf7\") pod \"ad76984a-1396-4144-8f61-664cd77cb5aa\" (UID: \"ad76984a-1396-4144-8f61-664cd77cb5aa\") " Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.375644 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-config\") pod \"ad76984a-1396-4144-8f61-664cd77cb5aa\" (UID: \"ad76984a-1396-4144-8f61-664cd77cb5aa\") " Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.375709 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-dns-svc\") pod \"ad76984a-1396-4144-8f61-664cd77cb5aa\" (UID: \"ad76984a-1396-4144-8f61-664cd77cb5aa\") " Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.375764 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-ovsdbserver-sb\") pod \"ad76984a-1396-4144-8f61-664cd77cb5aa\" (UID: \"ad76984a-1396-4144-8f61-664cd77cb5aa\") " Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.375880 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-ovsdbserver-nb\") pod \"ad76984a-1396-4144-8f61-664cd77cb5aa\" (UID: \"ad76984a-1396-4144-8f61-664cd77cb5aa\") " Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.383699 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.389479 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad76984a-1396-4144-8f61-664cd77cb5aa-kube-api-access-8hgf7" (OuterVolumeSpecName: "kube-api-access-8hgf7") pod "ad76984a-1396-4144-8f61-664cd77cb5aa" (UID: "ad76984a-1396-4144-8f61-664cd77cb5aa"). InnerVolumeSpecName "kube-api-access-8hgf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.419654 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="435bbb18-4b67-4b64-a922-f57717d8adc8" path="/var/lib/kubelet/pods/435bbb18-4b67-4b64-a922-f57717d8adc8/volumes" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.421471 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ad76984a-1396-4144-8f61-664cd77cb5aa" (UID: "ad76984a-1396-4144-8f61-664cd77cb5aa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.434797 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-config" (OuterVolumeSpecName: "config") pod "ad76984a-1396-4144-8f61-664cd77cb5aa" (UID: "ad76984a-1396-4144-8f61-664cd77cb5aa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.439618 4808 generic.go:334] "Generic (PLEG): container finished" podID="ad76984a-1396-4144-8f61-664cd77cb5aa" containerID="2c31739baa1672ba3aba3c1eab11c077378d920ea415d3964edf6753b64df227" exitCode=0 Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.439708 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-d44pp" event={"ID":"ad76984a-1396-4144-8f61-664cd77cb5aa","Type":"ContainerDied","Data":"2c31739baa1672ba3aba3c1eab11c077378d920ea415d3964edf6753b64df227"} Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.439744 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-d44pp" event={"ID":"ad76984a-1396-4144-8f61-664cd77cb5aa","Type":"ContainerDied","Data":"ca2b81ce2b5124d39e9933b428a4478504582263914a112d991421af411042bb"} Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.439766 4808 scope.go:117] "RemoveContainer" containerID="2c31739baa1672ba3aba3c1eab11c077378d920ea415d3964edf6753b64df227" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.439708 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b946d459c-d44pp" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.442428 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c16fcba0-e224-4b8f-a220-71184caeede4","Type":"ContainerStarted","Data":"18124b83073c3a542a8d914cb41ef4cb6e5710a9c65c7970f3cdcdaec177f5ac"} Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.443804 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ad76984a-1396-4144-8f61-664cd77cb5aa" (UID: "ad76984a-1396-4144-8f61-664cd77cb5aa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.450750 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ad76984a-1396-4144-8f61-664cd77cb5aa" (UID: "ad76984a-1396-4144-8f61-664cd77cb5aa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.474062 4808 scope.go:117] "RemoveContainer" containerID="e4ef75e980f4fffe5411e2b4df718abdfb17ca2b9a085c71bd5fbf1ded03e95a" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.477886 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.477913 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.477928 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.477939 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hgf7\" (UniqueName: \"kubernetes.io/projected/ad76984a-1396-4144-8f61-664cd77cb5aa-kube-api-access-8hgf7\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.477950 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad76984a-1396-4144-8f61-664cd77cb5aa-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.504547 4808 scope.go:117] "RemoveContainer" containerID="2c31739baa1672ba3aba3c1eab11c077378d920ea415d3964edf6753b64df227" Oct 02 16:58:25 crc kubenswrapper[4808]: E1002 16:58:25.504979 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c31739baa1672ba3aba3c1eab11c077378d920ea415d3964edf6753b64df227\": container with ID starting with 2c31739baa1672ba3aba3c1eab11c077378d920ea415d3964edf6753b64df227 not found: ID does not exist" containerID="2c31739baa1672ba3aba3c1eab11c077378d920ea415d3964edf6753b64df227" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.505009 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c31739baa1672ba3aba3c1eab11c077378d920ea415d3964edf6753b64df227"} err="failed to get container status \"2c31739baa1672ba3aba3c1eab11c077378d920ea415d3964edf6753b64df227\": rpc error: code = NotFound desc = could not find container \"2c31739baa1672ba3aba3c1eab11c077378d920ea415d3964edf6753b64df227\": container with ID starting with 2c31739baa1672ba3aba3c1eab11c077378d920ea415d3964edf6753b64df227 not found: ID does not exist" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.505029 4808 scope.go:117] "RemoveContainer" containerID="e4ef75e980f4fffe5411e2b4df718abdfb17ca2b9a085c71bd5fbf1ded03e95a" Oct 02 16:58:25 crc kubenswrapper[4808]: E1002 16:58:25.505210 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4ef75e980f4fffe5411e2b4df718abdfb17ca2b9a085c71bd5fbf1ded03e95a\": container with ID starting with e4ef75e980f4fffe5411e2b4df718abdfb17ca2b9a085c71bd5fbf1ded03e95a not found: ID does not exist" containerID="e4ef75e980f4fffe5411e2b4df718abdfb17ca2b9a085c71bd5fbf1ded03e95a" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.505242 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4ef75e980f4fffe5411e2b4df718abdfb17ca2b9a085c71bd5fbf1ded03e95a"} err="failed to get container status \"e4ef75e980f4fffe5411e2b4df718abdfb17ca2b9a085c71bd5fbf1ded03e95a\": rpc error: code = NotFound desc = could not find container \"e4ef75e980f4fffe5411e2b4df718abdfb17ca2b9a085c71bd5fbf1ded03e95a\": container with ID starting with e4ef75e980f4fffe5411e2b4df718abdfb17ca2b9a085c71bd5fbf1ded03e95a not found: ID does not exist" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.572848 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-54854ffb46-kktlh"] Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.684956 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-gb8s6"] Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.693343 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5b4f6b689c-zbtm2"] Oct 02 16:58:25 crc kubenswrapper[4808]: W1002 16:58:25.708325 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod695150b9_231f_4f3b_8e53_47b067a55025.slice/crio-28766d3ba056b8d1c37e6e496f491b0bcfeea55b0f37859728b764e558489aac WatchSource:0}: Error finding container 28766d3ba056b8d1c37e6e496f491b0bcfeea55b0f37859728b764e558489aac: Status 404 returned error can't find the container with id 28766d3ba056b8d1c37e6e496f491b0bcfeea55b0f37859728b764e558489aac Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.780738 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-l2fls" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.786310 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-798c55655d-t754s"] Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.800769 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-d44pp"] Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.815873 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-d44pp"] Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.883163 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bd29c\" (UniqueName: \"kubernetes.io/projected/b3d4166a-c265-4cb4-8481-32e6b0144997-kube-api-access-bd29c\") pod \"b3d4166a-c265-4cb4-8481-32e6b0144997\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.883270 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b3d4166a-c265-4cb4-8481-32e6b0144997-etc-machine-id\") pod \"b3d4166a-c265-4cb4-8481-32e6b0144997\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.883337 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-combined-ca-bundle\") pod \"b3d4166a-c265-4cb4-8481-32e6b0144997\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.883379 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-db-sync-config-data\") pod \"b3d4166a-c265-4cb4-8481-32e6b0144997\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.883402 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-config-data\") pod \"b3d4166a-c265-4cb4-8481-32e6b0144997\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.883430 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b3d4166a-c265-4cb4-8481-32e6b0144997-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b3d4166a-c265-4cb4-8481-32e6b0144997" (UID: "b3d4166a-c265-4cb4-8481-32e6b0144997"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.883478 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-scripts\") pod \"b3d4166a-c265-4cb4-8481-32e6b0144997\" (UID: \"b3d4166a-c265-4cb4-8481-32e6b0144997\") " Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.883834 4808 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b3d4166a-c265-4cb4-8481-32e6b0144997-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.887024 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-scripts" (OuterVolumeSpecName: "scripts") pod "b3d4166a-c265-4cb4-8481-32e6b0144997" (UID: "b3d4166a-c265-4cb4-8481-32e6b0144997"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.888637 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b3d4166a-c265-4cb4-8481-32e6b0144997" (UID: "b3d4166a-c265-4cb4-8481-32e6b0144997"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.889034 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3d4166a-c265-4cb4-8481-32e6b0144997-kube-api-access-bd29c" (OuterVolumeSpecName: "kube-api-access-bd29c") pod "b3d4166a-c265-4cb4-8481-32e6b0144997" (UID: "b3d4166a-c265-4cb4-8481-32e6b0144997"). InnerVolumeSpecName "kube-api-access-bd29c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.906805 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3d4166a-c265-4cb4-8481-32e6b0144997" (UID: "b3d4166a-c265-4cb4-8481-32e6b0144997"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.926360 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-config-data" (OuterVolumeSpecName: "config-data") pod "b3d4166a-c265-4cb4-8481-32e6b0144997" (UID: "b3d4166a-c265-4cb4-8481-32e6b0144997"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.985297 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.985336 4808 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.985348 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.985360 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3d4166a-c265-4cb4-8481-32e6b0144997-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:25 crc kubenswrapper[4808]: I1002 16:58:25.985371 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bd29c\" (UniqueName: \"kubernetes.io/projected/b3d4166a-c265-4cb4-8481-32e6b0144997-kube-api-access-bd29c\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.456361 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-l2fls" event={"ID":"b3d4166a-c265-4cb4-8481-32e6b0144997","Type":"ContainerDied","Data":"a955a03dc136394e57f80c0c6eeeb7b7a798bef982992e563dfae51f03c36bfe"} Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.456735 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a955a03dc136394e57f80c0c6eeeb7b7a798bef982992e563dfae51f03c36bfe" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.456619 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-l2fls" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.464998 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c16fcba0-e224-4b8f-a220-71184caeede4","Type":"ContainerStarted","Data":"2587f9f7ff50a6b9964bb34e849273fe79595b5a4f5831e89e480a2abe53d4ea"} Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.468189 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" event={"ID":"f36f9202-6bbb-404f-865d-206a87585936","Type":"ContainerStarted","Data":"b4314e8fe13dc3c27cf9721cf57dcd2c6ae928e2452dba0e4f61899033d8a6cd"} Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.473925 4808 generic.go:334] "Generic (PLEG): container finished" podID="e4ed4b3e-896b-4d66-b340-deecd1501807" containerID="3dcc2d158ec7a62a82c6540ee31a79912816be1b76256bc11f0f4256e2ca4e6d" exitCode=0 Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.474017 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" event={"ID":"e4ed4b3e-896b-4d66-b340-deecd1501807","Type":"ContainerDied","Data":"3dcc2d158ec7a62a82c6540ee31a79912816be1b76256bc11f0f4256e2ca4e6d"} Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.474062 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" event={"ID":"e4ed4b3e-896b-4d66-b340-deecd1501807","Type":"ContainerStarted","Data":"ff3c29ffd6f9dae44a62b89bad3d8b69d3ff3a4dca6ca0a6cccbaa1c8e19597b"} Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.512718 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-798c55655d-t754s" event={"ID":"28edd780-f606-40e4-a9a3-babce79c3f0f","Type":"ContainerStarted","Data":"ddb6929dc87352bd47c3e477cd0f5d8ef42ac3c2f630251e5c5634ec853b998e"} Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.512792 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.512805 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-798c55655d-t754s" event={"ID":"28edd780-f606-40e4-a9a3-babce79c3f0f","Type":"ContainerStarted","Data":"8e2fa9afe1086d68dafc20560b71896db7692c239f11b186ff876d372c2bcf8b"} Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.512816 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.519003 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b4f6b689c-zbtm2" event={"ID":"695150b9-231f-4f3b-8e53-47b067a55025","Type":"ContainerStarted","Data":"28766d3ba056b8d1c37e6e496f491b0bcfeea55b0f37859728b764e558489aac"} Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.538418 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-798c55655d-t754s" podStartSLOduration=2.538400014 podStartE2EDuration="2.538400014s" podCreationTimestamp="2025-10-02 16:58:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:58:26.526656295 +0000 UTC m=+1073.852185285" watchObservedRunningTime="2025-10-02 16:58:26.538400014 +0000 UTC m=+1073.863929014" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.676702 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 16:58:26 crc kubenswrapper[4808]: E1002 16:58:26.677293 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad76984a-1396-4144-8f61-664cd77cb5aa" containerName="dnsmasq-dns" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.677305 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad76984a-1396-4144-8f61-664cd77cb5aa" containerName="dnsmasq-dns" Oct 02 16:58:26 crc kubenswrapper[4808]: E1002 16:58:26.677329 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d4166a-c265-4cb4-8481-32e6b0144997" containerName="cinder-db-sync" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.677335 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d4166a-c265-4cb4-8481-32e6b0144997" containerName="cinder-db-sync" Oct 02 16:58:26 crc kubenswrapper[4808]: E1002 16:58:26.677352 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad76984a-1396-4144-8f61-664cd77cb5aa" containerName="init" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.677358 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad76984a-1396-4144-8f61-664cd77cb5aa" containerName="init" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.677519 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad76984a-1396-4144-8f61-664cd77cb5aa" containerName="dnsmasq-dns" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.677535 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3d4166a-c265-4cb4-8481-32e6b0144997" containerName="cinder-db-sync" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.678348 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.682549 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.682825 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.683054 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.683197 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-q2dqb" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.687043 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.738153 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-gb8s6"] Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.748094 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-g2j2k"] Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.755332 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.774566 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-g2j2k"] Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.804034 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-config-data\") pod \"cinder-scheduler-0\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.804077 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-scripts\") pod \"cinder-scheduler-0\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.804132 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.804173 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.804191 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpcrv\" (UniqueName: \"kubernetes.io/projected/2bf755e9-580a-497d-bb69-8f886fcb8dc4-kube-api-access-xpcrv\") pod \"cinder-scheduler-0\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.804216 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2bf755e9-580a-497d-bb69-8f886fcb8dc4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.839505 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.840904 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.845724 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.854495 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.909399 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-dns-svc\") pod \"dnsmasq-dns-6d97fcdd8f-g2j2k\" (UID: \"c166b259-1f16-4e18-8df1-a4a222d49897\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.909477 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.909503 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpcrv\" (UniqueName: \"kubernetes.io/projected/2bf755e9-580a-497d-bb69-8f886fcb8dc4-kube-api-access-xpcrv\") pod \"cinder-scheduler-0\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.909532 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2bf755e9-580a-497d-bb69-8f886fcb8dc4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.909548 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-ovsdbserver-sb\") pod \"dnsmasq-dns-6d97fcdd8f-g2j2k\" (UID: \"c166b259-1f16-4e18-8df1-a4a222d49897\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.909583 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-config\") pod \"dnsmasq-dns-6d97fcdd8f-g2j2k\" (UID: \"c166b259-1f16-4e18-8df1-a4a222d49897\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.909616 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-config-data\") pod \"cinder-scheduler-0\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.909634 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-scripts\") pod \"cinder-scheduler-0\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.909660 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-ovsdbserver-nb\") pod \"dnsmasq-dns-6d97fcdd8f-g2j2k\" (UID: \"c166b259-1f16-4e18-8df1-a4a222d49897\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.909677 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92ch8\" (UniqueName: \"kubernetes.io/projected/c166b259-1f16-4e18-8df1-a4a222d49897-kube-api-access-92ch8\") pod \"dnsmasq-dns-6d97fcdd8f-g2j2k\" (UID: \"c166b259-1f16-4e18-8df1-a4a222d49897\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.909715 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.910630 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2bf755e9-580a-497d-bb69-8f886fcb8dc4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.915986 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-scripts\") pod \"cinder-scheduler-0\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.916580 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-config-data\") pod \"cinder-scheduler-0\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.920665 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.935928 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:26 crc kubenswrapper[4808]: I1002 16:58:26.938907 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpcrv\" (UniqueName: \"kubernetes.io/projected/2bf755e9-580a-497d-bb69-8f886fcb8dc4-kube-api-access-xpcrv\") pod \"cinder-scheduler-0\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.004589 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.011385 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-scripts\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.011440 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-ovsdbserver-nb\") pod \"dnsmasq-dns-6d97fcdd8f-g2j2k\" (UID: \"c166b259-1f16-4e18-8df1-a4a222d49897\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.011465 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92ch8\" (UniqueName: \"kubernetes.io/projected/c166b259-1f16-4e18-8df1-a4a222d49897-kube-api-access-92ch8\") pod \"dnsmasq-dns-6d97fcdd8f-g2j2k\" (UID: \"c166b259-1f16-4e18-8df1-a4a222d49897\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.011488 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gzr9\" (UniqueName: \"kubernetes.io/projected/a5283df4-c936-4cb7-b6bf-db78d6cc8161-kube-api-access-4gzr9\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.011535 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-dns-svc\") pod \"dnsmasq-dns-6d97fcdd8f-g2j2k\" (UID: \"c166b259-1f16-4e18-8df1-a4a222d49897\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.011552 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5283df4-c936-4cb7-b6bf-db78d6cc8161-logs\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.011573 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.011595 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5283df4-c936-4cb7-b6bf-db78d6cc8161-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.011628 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-config-data-custom\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.011665 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-ovsdbserver-sb\") pod \"dnsmasq-dns-6d97fcdd8f-g2j2k\" (UID: \"c166b259-1f16-4e18-8df1-a4a222d49897\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.011703 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-config\") pod \"dnsmasq-dns-6d97fcdd8f-g2j2k\" (UID: \"c166b259-1f16-4e18-8df1-a4a222d49897\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.011723 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-config-data\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.012225 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-ovsdbserver-nb\") pod \"dnsmasq-dns-6d97fcdd8f-g2j2k\" (UID: \"c166b259-1f16-4e18-8df1-a4a222d49897\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.013090 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-config\") pod \"dnsmasq-dns-6d97fcdd8f-g2j2k\" (UID: \"c166b259-1f16-4e18-8df1-a4a222d49897\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.013116 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-dns-svc\") pod \"dnsmasq-dns-6d97fcdd8f-g2j2k\" (UID: \"c166b259-1f16-4e18-8df1-a4a222d49897\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.013513 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-ovsdbserver-sb\") pod \"dnsmasq-dns-6d97fcdd8f-g2j2k\" (UID: \"c166b259-1f16-4e18-8df1-a4a222d49897\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.039768 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92ch8\" (UniqueName: \"kubernetes.io/projected/c166b259-1f16-4e18-8df1-a4a222d49897-kube-api-access-92ch8\") pod \"dnsmasq-dns-6d97fcdd8f-g2j2k\" (UID: \"c166b259-1f16-4e18-8df1-a4a222d49897\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.109520 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.113662 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5283df4-c936-4cb7-b6bf-db78d6cc8161-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.113731 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-config-data-custom\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.113831 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-config-data\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.113873 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-scripts\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.113917 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gzr9\" (UniqueName: \"kubernetes.io/projected/a5283df4-c936-4cb7-b6bf-db78d6cc8161-kube-api-access-4gzr9\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.113970 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5283df4-c936-4cb7-b6bf-db78d6cc8161-logs\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.113993 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.116696 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5283df4-c936-4cb7-b6bf-db78d6cc8161-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.118225 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-config-data\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.118585 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.118783 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5283df4-c936-4cb7-b6bf-db78d6cc8161-logs\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.119524 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-scripts\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.127873 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-config-data-custom\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.133842 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gzr9\" (UniqueName: \"kubernetes.io/projected/a5283df4-c936-4cb7-b6bf-db78d6cc8161-kube-api-access-4gzr9\") pod \"cinder-api-0\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.205899 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.409399 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad76984a-1396-4144-8f61-664cd77cb5aa" path="/var/lib/kubelet/pods/ad76984a-1396-4144-8f61-664cd77cb5aa/volumes" Oct 02 16:58:27 crc kubenswrapper[4808]: I1002 16:58:27.532794 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-798c55655d-t754s" event={"ID":"28edd780-f606-40e4-a9a3-babce79c3f0f","Type":"ContainerStarted","Data":"d9203c1a421d591e83f7a8078f645d2b588113534092e579d107f92b6d697304"} Oct 02 16:58:28 crc kubenswrapper[4808]: I1002 16:58:28.715567 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 16:58:28 crc kubenswrapper[4808]: I1002 16:58:28.825741 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 16:58:28 crc kubenswrapper[4808]: I1002 16:58:28.902486 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-g2j2k"] Oct 02 16:58:29 crc kubenswrapper[4808]: I1002 16:58:29.553561 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a5283df4-c936-4cb7-b6bf-db78d6cc8161","Type":"ContainerStarted","Data":"3c7b4b649807d8f4b6b025f50cda4768e90001e92bf9131203513ec9c1ee029c"} Oct 02 16:58:29 crc kubenswrapper[4808]: I1002 16:58:29.554149 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a5283df4-c936-4cb7-b6bf-db78d6cc8161","Type":"ContainerStarted","Data":"65cdb72c41f54d0e955d28f7848d07c2888d83b37c88fd680085be479110a41f"} Oct 02 16:58:29 crc kubenswrapper[4808]: I1002 16:58:29.555751 4808 generic.go:334] "Generic (PLEG): container finished" podID="c166b259-1f16-4e18-8df1-a4a222d49897" containerID="728155e665696413326ab58d71546bb2627e8d71d48745322b77e76a0064a596" exitCode=0 Oct 02 16:58:29 crc kubenswrapper[4808]: I1002 16:58:29.555819 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" event={"ID":"c166b259-1f16-4e18-8df1-a4a222d49897","Type":"ContainerDied","Data":"728155e665696413326ab58d71546bb2627e8d71d48745322b77e76a0064a596"} Oct 02 16:58:29 crc kubenswrapper[4808]: I1002 16:58:29.555847 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" event={"ID":"c166b259-1f16-4e18-8df1-a4a222d49897","Type":"ContainerStarted","Data":"c58fbada1c5d03748625c8809ce46ab79b82807a863c677ec91824365e96ba5c"} Oct 02 16:58:29 crc kubenswrapper[4808]: I1002 16:58:29.557127 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b4f6b689c-zbtm2" event={"ID":"695150b9-231f-4f3b-8e53-47b067a55025","Type":"ContainerStarted","Data":"d44e9206e8ca85660505aad52f76b0982fa5d07189ad7177bc4c5f9fec85b43c"} Oct 02 16:58:29 crc kubenswrapper[4808]: I1002 16:58:29.557155 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b4f6b689c-zbtm2" event={"ID":"695150b9-231f-4f3b-8e53-47b067a55025","Type":"ContainerStarted","Data":"816421917e68dda6fbbdc1a643ea3b634d564e1e34488b183760ea4b7e4ae451"} Oct 02 16:58:29 crc kubenswrapper[4808]: I1002 16:58:29.565414 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c16fcba0-e224-4b8f-a220-71184caeede4","Type":"ContainerStarted","Data":"85f033100798bc2c00717d786e992f63420eb42ea2853f1627165cf4d08e2a06"} Oct 02 16:58:29 crc kubenswrapper[4808]: I1002 16:58:29.565532 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c16fcba0-e224-4b8f-a220-71184caeede4","Type":"ContainerStarted","Data":"438653f745c160c504df307e4d593d42e716a38f68e29354622a4c8a163a689c"} Oct 02 16:58:29 crc kubenswrapper[4808]: I1002 16:58:29.567184 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" event={"ID":"f36f9202-6bbb-404f-865d-206a87585936","Type":"ContainerStarted","Data":"df9c586b742b9f3e383a3acc51dd11bd6baf6edfb6c8298442a91d6973a4b854"} Oct 02 16:58:29 crc kubenswrapper[4808]: I1002 16:58:29.567204 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" event={"ID":"f36f9202-6bbb-404f-865d-206a87585936","Type":"ContainerStarted","Data":"84355841f73955eaaf06e5b967cef32eab576b8baa7d3397caade67748fe6c5b"} Oct 02 16:58:29 crc kubenswrapper[4808]: I1002 16:58:29.569576 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2bf755e9-580a-497d-bb69-8f886fcb8dc4","Type":"ContainerStarted","Data":"95913a43205d00578535fc3ca12973f45a2f8733ef383aa8ca295486366889a4"} Oct 02 16:58:29 crc kubenswrapper[4808]: I1002 16:58:29.611424 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" event={"ID":"e4ed4b3e-896b-4d66-b340-deecd1501807","Type":"ContainerStarted","Data":"7331ca11f738900578207cc5c4995d7285e2c47dc10f4590451248f41a9325f4"} Oct 02 16:58:29 crc kubenswrapper[4808]: I1002 16:58:29.611614 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" podUID="e4ed4b3e-896b-4d66-b340-deecd1501807" containerName="dnsmasq-dns" containerID="cri-o://7331ca11f738900578207cc5c4995d7285e2c47dc10f4590451248f41a9325f4" gracePeriod=10 Oct 02 16:58:29 crc kubenswrapper[4808]: I1002 16:58:29.611870 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" Oct 02 16:58:29 crc kubenswrapper[4808]: I1002 16:58:29.616830 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-54854ffb46-kktlh" podStartSLOduration=2.971646095 podStartE2EDuration="5.616810423s" podCreationTimestamp="2025-10-02 16:58:24 +0000 UTC" firstStartedPulling="2025-10-02 16:58:25.586486742 +0000 UTC m=+1072.912015762" lastFinishedPulling="2025-10-02 16:58:28.23165109 +0000 UTC m=+1075.557180090" observedRunningTime="2025-10-02 16:58:29.603980804 +0000 UTC m=+1076.929509824" watchObservedRunningTime="2025-10-02 16:58:29.616810423 +0000 UTC m=+1076.942339423" Oct 02 16:58:29 crc kubenswrapper[4808]: I1002 16:58:29.639431 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5b4f6b689c-zbtm2" podStartSLOduration=3.069857761 podStartE2EDuration="5.639402276s" podCreationTimestamp="2025-10-02 16:58:24 +0000 UTC" firstStartedPulling="2025-10-02 16:58:25.715176655 +0000 UTC m=+1073.040705645" lastFinishedPulling="2025-10-02 16:58:28.28472115 +0000 UTC m=+1075.610250160" observedRunningTime="2025-10-02 16:58:29.624734668 +0000 UTC m=+1076.950263668" watchObservedRunningTime="2025-10-02 16:58:29.639402276 +0000 UTC m=+1076.964931276" Oct 02 16:58:29 crc kubenswrapper[4808]: I1002 16:58:29.658936 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" podStartSLOduration=5.658911856 podStartE2EDuration="5.658911856s" podCreationTimestamp="2025-10-02 16:58:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:58:29.650194519 +0000 UTC m=+1076.975723529" watchObservedRunningTime="2025-10-02 16:58:29.658911856 +0000 UTC m=+1076.984440856" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.203021 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.302954 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-dns-svc\") pod \"e4ed4b3e-896b-4d66-b340-deecd1501807\" (UID: \"e4ed4b3e-896b-4d66-b340-deecd1501807\") " Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.303019 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-ovsdbserver-nb\") pod \"e4ed4b3e-896b-4d66-b340-deecd1501807\" (UID: \"e4ed4b3e-896b-4d66-b340-deecd1501807\") " Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.303043 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-config\") pod \"e4ed4b3e-896b-4d66-b340-deecd1501807\" (UID: \"e4ed4b3e-896b-4d66-b340-deecd1501807\") " Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.303152 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-ovsdbserver-sb\") pod \"e4ed4b3e-896b-4d66-b340-deecd1501807\" (UID: \"e4ed4b3e-896b-4d66-b340-deecd1501807\") " Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.303184 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsc5h\" (UniqueName: \"kubernetes.io/projected/e4ed4b3e-896b-4d66-b340-deecd1501807-kube-api-access-gsc5h\") pod \"e4ed4b3e-896b-4d66-b340-deecd1501807\" (UID: \"e4ed4b3e-896b-4d66-b340-deecd1501807\") " Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.316492 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4ed4b3e-896b-4d66-b340-deecd1501807-kube-api-access-gsc5h" (OuterVolumeSpecName: "kube-api-access-gsc5h") pod "e4ed4b3e-896b-4d66-b340-deecd1501807" (UID: "e4ed4b3e-896b-4d66-b340-deecd1501807"). InnerVolumeSpecName "kube-api-access-gsc5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.358546 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e4ed4b3e-896b-4d66-b340-deecd1501807" (UID: "e4ed4b3e-896b-4d66-b340-deecd1501807"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.369530 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-config" (OuterVolumeSpecName: "config") pod "e4ed4b3e-896b-4d66-b340-deecd1501807" (UID: "e4ed4b3e-896b-4d66-b340-deecd1501807"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.375726 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e4ed4b3e-896b-4d66-b340-deecd1501807" (UID: "e4ed4b3e-896b-4d66-b340-deecd1501807"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.381090 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e4ed4b3e-896b-4d66-b340-deecd1501807" (UID: "e4ed4b3e-896b-4d66-b340-deecd1501807"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.405049 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.405081 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.405091 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.405102 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e4ed4b3e-896b-4d66-b340-deecd1501807-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.405112 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsc5h\" (UniqueName: \"kubernetes.io/projected/e4ed4b3e-896b-4d66-b340-deecd1501807-kube-api-access-gsc5h\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.622864 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2bf755e9-580a-497d-bb69-8f886fcb8dc4","Type":"ContainerStarted","Data":"d9a695a9740bee4636c6be6d2b8147c5bc18a38a613d89d626afbf205a7830c3"} Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.627435 4808 generic.go:334] "Generic (PLEG): container finished" podID="e4ed4b3e-896b-4d66-b340-deecd1501807" containerID="7331ca11f738900578207cc5c4995d7285e2c47dc10f4590451248f41a9325f4" exitCode=0 Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.627490 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.627539 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" event={"ID":"e4ed4b3e-896b-4d66-b340-deecd1501807","Type":"ContainerDied","Data":"7331ca11f738900578207cc5c4995d7285e2c47dc10f4590451248f41a9325f4"} Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.627630 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb684768f-gb8s6" event={"ID":"e4ed4b3e-896b-4d66-b340-deecd1501807","Type":"ContainerDied","Data":"ff3c29ffd6f9dae44a62b89bad3d8b69d3ff3a4dca6ca0a6cccbaa1c8e19597b"} Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.627678 4808 scope.go:117] "RemoveContainer" containerID="7331ca11f738900578207cc5c4995d7285e2c47dc10f4590451248f41a9325f4" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.637797 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a5283df4-c936-4cb7-b6bf-db78d6cc8161","Type":"ContainerStarted","Data":"1a944d09177bd07227870ec39de8e7cb969323434734daa628b0cb2e3605d581"} Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.637882 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.645315 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" event={"ID":"c166b259-1f16-4e18-8df1-a4a222d49897","Type":"ContainerStarted","Data":"39095e68ee42cf7a1370c4cbb6399805c01b1864ac520be1c9a337fb1d065ba2"} Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.661303 4808 scope.go:117] "RemoveContainer" containerID="3dcc2d158ec7a62a82c6540ee31a79912816be1b76256bc11f0f4256e2ca4e6d" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.662123 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.662075999 podStartE2EDuration="4.662075999s" podCreationTimestamp="2025-10-02 16:58:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:58:30.65108213 +0000 UTC m=+1077.976611140" watchObservedRunningTime="2025-10-02 16:58:30.662075999 +0000 UTC m=+1077.987604999" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.675491 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" podStartSLOduration=4.675476992 podStartE2EDuration="4.675476992s" podCreationTimestamp="2025-10-02 16:58:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:58:30.67427775 +0000 UTC m=+1077.999806760" watchObservedRunningTime="2025-10-02 16:58:30.675476992 +0000 UTC m=+1078.001006002" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.709572 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-gb8s6"] Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.711543 4808 scope.go:117] "RemoveContainer" containerID="7331ca11f738900578207cc5c4995d7285e2c47dc10f4590451248f41a9325f4" Oct 02 16:58:30 crc kubenswrapper[4808]: E1002 16:58:30.714457 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7331ca11f738900578207cc5c4995d7285e2c47dc10f4590451248f41a9325f4\": container with ID starting with 7331ca11f738900578207cc5c4995d7285e2c47dc10f4590451248f41a9325f4 not found: ID does not exist" containerID="7331ca11f738900578207cc5c4995d7285e2c47dc10f4590451248f41a9325f4" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.714505 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7331ca11f738900578207cc5c4995d7285e2c47dc10f4590451248f41a9325f4"} err="failed to get container status \"7331ca11f738900578207cc5c4995d7285e2c47dc10f4590451248f41a9325f4\": rpc error: code = NotFound desc = could not find container \"7331ca11f738900578207cc5c4995d7285e2c47dc10f4590451248f41a9325f4\": container with ID starting with 7331ca11f738900578207cc5c4995d7285e2c47dc10f4590451248f41a9325f4 not found: ID does not exist" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.714535 4808 scope.go:117] "RemoveContainer" containerID="3dcc2d158ec7a62a82c6540ee31a79912816be1b76256bc11f0f4256e2ca4e6d" Oct 02 16:58:30 crc kubenswrapper[4808]: E1002 16:58:30.715210 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dcc2d158ec7a62a82c6540ee31a79912816be1b76256bc11f0f4256e2ca4e6d\": container with ID starting with 3dcc2d158ec7a62a82c6540ee31a79912816be1b76256bc11f0f4256e2ca4e6d not found: ID does not exist" containerID="3dcc2d158ec7a62a82c6540ee31a79912816be1b76256bc11f0f4256e2ca4e6d" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.715263 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dcc2d158ec7a62a82c6540ee31a79912816be1b76256bc11f0f4256e2ca4e6d"} err="failed to get container status \"3dcc2d158ec7a62a82c6540ee31a79912816be1b76256bc11f0f4256e2ca4e6d\": rpc error: code = NotFound desc = could not find container \"3dcc2d158ec7a62a82c6540ee31a79912816be1b76256bc11f0f4256e2ca4e6d\": container with ID starting with 3dcc2d158ec7a62a82c6540ee31a79912816be1b76256bc11f0f4256e2ca4e6d not found: ID does not exist" Oct 02 16:58:30 crc kubenswrapper[4808]: I1002 16:58:30.722296 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-gb8s6"] Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.415046 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4ed4b3e-896b-4d66-b340-deecd1501807" path="/var/lib/kubelet/pods/e4ed4b3e-896b-4d66-b340-deecd1501807/volumes" Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.448043 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.654685 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2bf755e9-580a-497d-bb69-8f886fcb8dc4","Type":"ContainerStarted","Data":"20d11cfaabe7fd0fee6156d616541b3adb397d5a41ee987463fd71404b28fc08"} Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.657923 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c16fcba0-e224-4b8f-a220-71184caeede4","Type":"ContainerStarted","Data":"c488da3f08e05efc006ef52801722d928ad4f2a86704dfb0153b85eafaa838f6"} Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.658399 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.680797 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.604124526 podStartE2EDuration="5.680781704s" podCreationTimestamp="2025-10-02 16:58:26 +0000 UTC" firstStartedPulling="2025-10-02 16:58:28.743739851 +0000 UTC m=+1076.069268851" lastFinishedPulling="2025-10-02 16:58:29.820397029 +0000 UTC m=+1077.145926029" observedRunningTime="2025-10-02 16:58:31.677534216 +0000 UTC m=+1079.003063216" watchObservedRunningTime="2025-10-02 16:58:31.680781704 +0000 UTC m=+1079.006310694" Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.708151 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.336588154 podStartE2EDuration="7.708136446s" podCreationTimestamp="2025-10-02 16:58:24 +0000 UTC" firstStartedPulling="2025-10-02 16:58:25.392522986 +0000 UTC m=+1072.718051976" lastFinishedPulling="2025-10-02 16:58:30.764071268 +0000 UTC m=+1078.089600268" observedRunningTime="2025-10-02 16:58:31.703581923 +0000 UTC m=+1079.029110923" watchObservedRunningTime="2025-10-02 16:58:31.708136446 +0000 UTC m=+1079.033665446" Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.889650 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5684d76658-g9f5f"] Oct 02 16:58:31 crc kubenswrapper[4808]: E1002 16:58:31.889950 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4ed4b3e-896b-4d66-b340-deecd1501807" containerName="init" Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.889966 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4ed4b3e-896b-4d66-b340-deecd1501807" containerName="init" Oct 02 16:58:31 crc kubenswrapper[4808]: E1002 16:58:31.889997 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4ed4b3e-896b-4d66-b340-deecd1501807" containerName="dnsmasq-dns" Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.890003 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4ed4b3e-896b-4d66-b340-deecd1501807" containerName="dnsmasq-dns" Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.890147 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4ed4b3e-896b-4d66-b340-deecd1501807" containerName="dnsmasq-dns" Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.890960 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.893481 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.894592 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.910609 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5684d76658-g9f5f"] Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.936450 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-internal-tls-certs\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.936499 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-logs\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.936523 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-config-data-custom\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.936547 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-combined-ca-bundle\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.936572 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx5jb\" (UniqueName: \"kubernetes.io/projected/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-kube-api-access-wx5jb\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.936593 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-public-tls-certs\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:31 crc kubenswrapper[4808]: I1002 16:58:31.936772 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-config-data\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:32 crc kubenswrapper[4808]: I1002 16:58:32.004905 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 16:58:32 crc kubenswrapper[4808]: I1002 16:58:32.038159 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-config-data\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:32 crc kubenswrapper[4808]: I1002 16:58:32.038343 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-internal-tls-certs\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:32 crc kubenswrapper[4808]: I1002 16:58:32.038380 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-logs\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:32 crc kubenswrapper[4808]: I1002 16:58:32.038414 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-config-data-custom\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:32 crc kubenswrapper[4808]: I1002 16:58:32.038444 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-combined-ca-bundle\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:32 crc kubenswrapper[4808]: I1002 16:58:32.038487 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx5jb\" (UniqueName: \"kubernetes.io/projected/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-kube-api-access-wx5jb\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:32 crc kubenswrapper[4808]: I1002 16:58:32.038513 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-public-tls-certs\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:32 crc kubenswrapper[4808]: I1002 16:58:32.038916 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-logs\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:32 crc kubenswrapper[4808]: I1002 16:58:32.042524 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-internal-tls-certs\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:32 crc kubenswrapper[4808]: I1002 16:58:32.044780 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-config-data-custom\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:32 crc kubenswrapper[4808]: I1002 16:58:32.044961 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-public-tls-certs\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:32 crc kubenswrapper[4808]: I1002 16:58:32.045173 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-combined-ca-bundle\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:32 crc kubenswrapper[4808]: I1002 16:58:32.046357 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-config-data\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:32 crc kubenswrapper[4808]: I1002 16:58:32.062288 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx5jb\" (UniqueName: \"kubernetes.io/projected/f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9-kube-api-access-wx5jb\") pod \"barbican-api-5684d76658-g9f5f\" (UID: \"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9\") " pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:32 crc kubenswrapper[4808]: I1002 16:58:32.206178 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:32 crc kubenswrapper[4808]: I1002 16:58:32.666860 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a5283df4-c936-4cb7-b6bf-db78d6cc8161" containerName="cinder-api-log" containerID="cri-o://3c7b4b649807d8f4b6b025f50cda4768e90001e92bf9131203513ec9c1ee029c" gracePeriod=30 Oct 02 16:58:32 crc kubenswrapper[4808]: I1002 16:58:32.666904 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a5283df4-c936-4cb7-b6bf-db78d6cc8161" containerName="cinder-api" containerID="cri-o://1a944d09177bd07227870ec39de8e7cb969323434734daa628b0cb2e3605d581" gracePeriod=30 Oct 02 16:58:32 crc kubenswrapper[4808]: I1002 16:58:32.667938 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 16:58:32 crc kubenswrapper[4808]: I1002 16:58:32.760909 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5684d76658-g9f5f"] Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.295413 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.364286 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-config-data-custom\") pod \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.364394 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5283df4-c936-4cb7-b6bf-db78d6cc8161-etc-machine-id\") pod \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.364502 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a5283df4-c936-4cb7-b6bf-db78d6cc8161-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a5283df4-c936-4cb7-b6bf-db78d6cc8161" (UID: "a5283df4-c936-4cb7-b6bf-db78d6cc8161"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.364536 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gzr9\" (UniqueName: \"kubernetes.io/projected/a5283df4-c936-4cb7-b6bf-db78d6cc8161-kube-api-access-4gzr9\") pod \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.364570 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-scripts\") pod \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.364623 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-config-data\") pod \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.364666 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-combined-ca-bundle\") pod \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.365147 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5283df4-c936-4cb7-b6bf-db78d6cc8161-logs\") pod \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\" (UID: \"a5283df4-c936-4cb7-b6bf-db78d6cc8161\") " Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.365544 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5283df4-c936-4cb7-b6bf-db78d6cc8161-logs" (OuterVolumeSpecName: "logs") pod "a5283df4-c936-4cb7-b6bf-db78d6cc8161" (UID: "a5283df4-c936-4cb7-b6bf-db78d6cc8161"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.365576 4808 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5283df4-c936-4cb7-b6bf-db78d6cc8161-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.369190 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-scripts" (OuterVolumeSpecName: "scripts") pod "a5283df4-c936-4cb7-b6bf-db78d6cc8161" (UID: "a5283df4-c936-4cb7-b6bf-db78d6cc8161"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.370330 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5283df4-c936-4cb7-b6bf-db78d6cc8161-kube-api-access-4gzr9" (OuterVolumeSpecName: "kube-api-access-4gzr9") pod "a5283df4-c936-4cb7-b6bf-db78d6cc8161" (UID: "a5283df4-c936-4cb7-b6bf-db78d6cc8161"). InnerVolumeSpecName "kube-api-access-4gzr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.370421 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a5283df4-c936-4cb7-b6bf-db78d6cc8161" (UID: "a5283df4-c936-4cb7-b6bf-db78d6cc8161"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.393867 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5283df4-c936-4cb7-b6bf-db78d6cc8161" (UID: "a5283df4-c936-4cb7-b6bf-db78d6cc8161"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.464266 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-config-data" (OuterVolumeSpecName: "config-data") pod "a5283df4-c936-4cb7-b6bf-db78d6cc8161" (UID: "a5283df4-c936-4cb7-b6bf-db78d6cc8161"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.467589 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5283df4-c936-4cb7-b6bf-db78d6cc8161-logs\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.467629 4808 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.467667 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gzr9\" (UniqueName: \"kubernetes.io/projected/a5283df4-c936-4cb7-b6bf-db78d6cc8161-kube-api-access-4gzr9\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.467683 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.467694 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.467707 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5283df4-c936-4cb7-b6bf-db78d6cc8161-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.678286 4808 generic.go:334] "Generic (PLEG): container finished" podID="a5283df4-c936-4cb7-b6bf-db78d6cc8161" containerID="1a944d09177bd07227870ec39de8e7cb969323434734daa628b0cb2e3605d581" exitCode=0 Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.678346 4808 generic.go:334] "Generic (PLEG): container finished" podID="a5283df4-c936-4cb7-b6bf-db78d6cc8161" containerID="3c7b4b649807d8f4b6b025f50cda4768e90001e92bf9131203513ec9c1ee029c" exitCode=143 Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.678367 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.678363 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a5283df4-c936-4cb7-b6bf-db78d6cc8161","Type":"ContainerDied","Data":"1a944d09177bd07227870ec39de8e7cb969323434734daa628b0cb2e3605d581"} Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.678499 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a5283df4-c936-4cb7-b6bf-db78d6cc8161","Type":"ContainerDied","Data":"3c7b4b649807d8f4b6b025f50cda4768e90001e92bf9131203513ec9c1ee029c"} Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.678515 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a5283df4-c936-4cb7-b6bf-db78d6cc8161","Type":"ContainerDied","Data":"65cdb72c41f54d0e955d28f7848d07c2888d83b37c88fd680085be479110a41f"} Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.678532 4808 scope.go:117] "RemoveContainer" containerID="1a944d09177bd07227870ec39de8e7cb969323434734daa628b0cb2e3605d581" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.683706 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5684d76658-g9f5f" event={"ID":"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9","Type":"ContainerStarted","Data":"b5d8be5de28e9def56f40c3b5bcc696f2152f7fb3f48933289670e8ab5a5299e"} Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.683755 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5684d76658-g9f5f" event={"ID":"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9","Type":"ContainerStarted","Data":"3ab203c678b57d541f4565ca27cfdc3e80a5dec8462345e5a9591c8578ddca92"} Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.683777 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5684d76658-g9f5f" event={"ID":"f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9","Type":"ContainerStarted","Data":"44dc258fcdf994fa8a2c41e71ab427fe689e26c253e1171dc707f5915a73bea0"} Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.721888 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5684d76658-g9f5f" podStartSLOduration=2.7218672919999998 podStartE2EDuration="2.721867292s" podCreationTimestamp="2025-10-02 16:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:58:33.715944452 +0000 UTC m=+1081.041473452" watchObservedRunningTime="2025-10-02 16:58:33.721867292 +0000 UTC m=+1081.047396312" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.732283 4808 scope.go:117] "RemoveContainer" containerID="3c7b4b649807d8f4b6b025f50cda4768e90001e92bf9131203513ec9c1ee029c" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.743548 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.752171 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.761669 4808 scope.go:117] "RemoveContainer" containerID="1a944d09177bd07227870ec39de8e7cb969323434734daa628b0cb2e3605d581" Oct 02 16:58:33 crc kubenswrapper[4808]: E1002 16:58:33.762269 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a944d09177bd07227870ec39de8e7cb969323434734daa628b0cb2e3605d581\": container with ID starting with 1a944d09177bd07227870ec39de8e7cb969323434734daa628b0cb2e3605d581 not found: ID does not exist" containerID="1a944d09177bd07227870ec39de8e7cb969323434734daa628b0cb2e3605d581" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.762348 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a944d09177bd07227870ec39de8e7cb969323434734daa628b0cb2e3605d581"} err="failed to get container status \"1a944d09177bd07227870ec39de8e7cb969323434734daa628b0cb2e3605d581\": rpc error: code = NotFound desc = could not find container \"1a944d09177bd07227870ec39de8e7cb969323434734daa628b0cb2e3605d581\": container with ID starting with 1a944d09177bd07227870ec39de8e7cb969323434734daa628b0cb2e3605d581 not found: ID does not exist" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.762393 4808 scope.go:117] "RemoveContainer" containerID="3c7b4b649807d8f4b6b025f50cda4768e90001e92bf9131203513ec9c1ee029c" Oct 02 16:58:33 crc kubenswrapper[4808]: E1002 16:58:33.764391 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c7b4b649807d8f4b6b025f50cda4768e90001e92bf9131203513ec9c1ee029c\": container with ID starting with 3c7b4b649807d8f4b6b025f50cda4768e90001e92bf9131203513ec9c1ee029c not found: ID does not exist" containerID="3c7b4b649807d8f4b6b025f50cda4768e90001e92bf9131203513ec9c1ee029c" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.764452 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c7b4b649807d8f4b6b025f50cda4768e90001e92bf9131203513ec9c1ee029c"} err="failed to get container status \"3c7b4b649807d8f4b6b025f50cda4768e90001e92bf9131203513ec9c1ee029c\": rpc error: code = NotFound desc = could not find container \"3c7b4b649807d8f4b6b025f50cda4768e90001e92bf9131203513ec9c1ee029c\": container with ID starting with 3c7b4b649807d8f4b6b025f50cda4768e90001e92bf9131203513ec9c1ee029c not found: ID does not exist" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.764487 4808 scope.go:117] "RemoveContainer" containerID="1a944d09177bd07227870ec39de8e7cb969323434734daa628b0cb2e3605d581" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.765440 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a944d09177bd07227870ec39de8e7cb969323434734daa628b0cb2e3605d581"} err="failed to get container status \"1a944d09177bd07227870ec39de8e7cb969323434734daa628b0cb2e3605d581\": rpc error: code = NotFound desc = could not find container \"1a944d09177bd07227870ec39de8e7cb969323434734daa628b0cb2e3605d581\": container with ID starting with 1a944d09177bd07227870ec39de8e7cb969323434734daa628b0cb2e3605d581 not found: ID does not exist" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.765525 4808 scope.go:117] "RemoveContainer" containerID="3c7b4b649807d8f4b6b025f50cda4768e90001e92bf9131203513ec9c1ee029c" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.766045 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c7b4b649807d8f4b6b025f50cda4768e90001e92bf9131203513ec9c1ee029c"} err="failed to get container status \"3c7b4b649807d8f4b6b025f50cda4768e90001e92bf9131203513ec9c1ee029c\": rpc error: code = NotFound desc = could not find container \"3c7b4b649807d8f4b6b025f50cda4768e90001e92bf9131203513ec9c1ee029c\": container with ID starting with 3c7b4b649807d8f4b6b025f50cda4768e90001e92bf9131203513ec9c1ee029c not found: ID does not exist" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.770950 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 16:58:33 crc kubenswrapper[4808]: E1002 16:58:33.771542 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5283df4-c936-4cb7-b6bf-db78d6cc8161" containerName="cinder-api-log" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.771573 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5283df4-c936-4cb7-b6bf-db78d6cc8161" containerName="cinder-api-log" Oct 02 16:58:33 crc kubenswrapper[4808]: E1002 16:58:33.771616 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5283df4-c936-4cb7-b6bf-db78d6cc8161" containerName="cinder-api" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.771630 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5283df4-c936-4cb7-b6bf-db78d6cc8161" containerName="cinder-api" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.771932 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5283df4-c936-4cb7-b6bf-db78d6cc8161" containerName="cinder-api-log" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.771974 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5283df4-c936-4cb7-b6bf-db78d6cc8161" containerName="cinder-api" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.773502 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.777154 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.777390 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.779943 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.787249 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.878401 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7723b385-3e7c-4c10-bf48-f2213ce31f16-logs\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.878445 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7723b385-3e7c-4c10-bf48-f2213ce31f16-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.878470 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w72zm\" (UniqueName: \"kubernetes.io/projected/7723b385-3e7c-4c10-bf48-f2213ce31f16-kube-api-access-w72zm\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.878518 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7723b385-3e7c-4c10-bf48-f2213ce31f16-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.878535 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7723b385-3e7c-4c10-bf48-f2213ce31f16-scripts\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.878557 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7723b385-3e7c-4c10-bf48-f2213ce31f16-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.878570 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7723b385-3e7c-4c10-bf48-f2213ce31f16-config-data\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.878605 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7723b385-3e7c-4c10-bf48-f2213ce31f16-config-data-custom\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.878672 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7723b385-3e7c-4c10-bf48-f2213ce31f16-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.980081 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7723b385-3e7c-4c10-bf48-f2213ce31f16-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.980163 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7723b385-3e7c-4c10-bf48-f2213ce31f16-logs\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.980182 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7723b385-3e7c-4c10-bf48-f2213ce31f16-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.980204 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w72zm\" (UniqueName: \"kubernetes.io/projected/7723b385-3e7c-4c10-bf48-f2213ce31f16-kube-api-access-w72zm\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.980268 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7723b385-3e7c-4c10-bf48-f2213ce31f16-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.980284 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7723b385-3e7c-4c10-bf48-f2213ce31f16-scripts\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.980308 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7723b385-3e7c-4c10-bf48-f2213ce31f16-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.980321 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7723b385-3e7c-4c10-bf48-f2213ce31f16-config-data\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.980349 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7723b385-3e7c-4c10-bf48-f2213ce31f16-config-data-custom\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.981519 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7723b385-3e7c-4c10-bf48-f2213ce31f16-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.981605 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7723b385-3e7c-4c10-bf48-f2213ce31f16-logs\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.984503 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7723b385-3e7c-4c10-bf48-f2213ce31f16-config-data-custom\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.984931 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7723b385-3e7c-4c10-bf48-f2213ce31f16-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.985130 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7723b385-3e7c-4c10-bf48-f2213ce31f16-config-data\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.985512 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7723b385-3e7c-4c10-bf48-f2213ce31f16-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.985913 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7723b385-3e7c-4c10-bf48-f2213ce31f16-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.986341 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7723b385-3e7c-4c10-bf48-f2213ce31f16-scripts\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:33 crc kubenswrapper[4808]: I1002 16:58:33.996113 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w72zm\" (UniqueName: \"kubernetes.io/projected/7723b385-3e7c-4c10-bf48-f2213ce31f16-kube-api-access-w72zm\") pod \"cinder-api-0\" (UID: \"7723b385-3e7c-4c10-bf48-f2213ce31f16\") " pod="openstack/cinder-api-0" Oct 02 16:58:34 crc kubenswrapper[4808]: I1002 16:58:34.115249 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 16:58:34 crc kubenswrapper[4808]: I1002 16:58:34.591110 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 16:58:34 crc kubenswrapper[4808]: W1002 16:58:34.595625 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7723b385_3e7c_4c10_bf48_f2213ce31f16.slice/crio-3288754d9f1df751a8e7ddd51a94735495f3595569e78ba4dfb5108c27af783d WatchSource:0}: Error finding container 3288754d9f1df751a8e7ddd51a94735495f3595569e78ba4dfb5108c27af783d: Status 404 returned error can't find the container with id 3288754d9f1df751a8e7ddd51a94735495f3595569e78ba4dfb5108c27af783d Oct 02 16:58:34 crc kubenswrapper[4808]: I1002 16:58:34.698973 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7723b385-3e7c-4c10-bf48-f2213ce31f16","Type":"ContainerStarted","Data":"3288754d9f1df751a8e7ddd51a94735495f3595569e78ba4dfb5108c27af783d"} Oct 02 16:58:34 crc kubenswrapper[4808]: I1002 16:58:34.699095 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:34 crc kubenswrapper[4808]: I1002 16:58:34.699159 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:35 crc kubenswrapper[4808]: I1002 16:58:35.413503 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5283df4-c936-4cb7-b6bf-db78d6cc8161" path="/var/lib/kubelet/pods/a5283df4-c936-4cb7-b6bf-db78d6cc8161/volumes" Oct 02 16:58:35 crc kubenswrapper[4808]: I1002 16:58:35.711060 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7723b385-3e7c-4c10-bf48-f2213ce31f16","Type":"ContainerStarted","Data":"a65ddaa7aa5b843926de7cfefd0deec83f4b3dd0db763e75dbbfdaef81a479e7"} Oct 02 16:58:36 crc kubenswrapper[4808]: I1002 16:58:36.334161 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:36 crc kubenswrapper[4808]: I1002 16:58:36.382902 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:36 crc kubenswrapper[4808]: I1002 16:58:36.723067 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7723b385-3e7c-4c10-bf48-f2213ce31f16","Type":"ContainerStarted","Data":"43c823239ad372a0153e06c2385eb5aaed12b44cb60b357f532438fc83de896a"} Oct 02 16:58:36 crc kubenswrapper[4808]: I1002 16:58:36.757858 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.757841109 podStartE2EDuration="3.757841109s" podCreationTimestamp="2025-10-02 16:58:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:58:36.753838001 +0000 UTC m=+1084.079367021" watchObservedRunningTime="2025-10-02 16:58:36.757841109 +0000 UTC m=+1084.083370109" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.112466 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.185933 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-kkks2"] Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.186192 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" podUID="92a07292-975e-492a-97da-f713cfb5dbce" containerName="dnsmasq-dns" containerID="cri-o://b398c4c8f8e515341ac81d5c11e1b88348a2cff349306d7299c715065f1530cb" gracePeriod=10 Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.258685 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.331879 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.641442 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.731528 4808 generic.go:334] "Generic (PLEG): container finished" podID="92a07292-975e-492a-97da-f713cfb5dbce" containerID="b398c4c8f8e515341ac81d5c11e1b88348a2cff349306d7299c715065f1530cb" exitCode=0 Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.732922 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="2bf755e9-580a-497d-bb69-8f886fcb8dc4" containerName="cinder-scheduler" containerID="cri-o://d9a695a9740bee4636c6be6d2b8147c5bc18a38a613d89d626afbf205a7830c3" gracePeriod=30 Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.733363 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.733826 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" event={"ID":"92a07292-975e-492a-97da-f713cfb5dbce","Type":"ContainerDied","Data":"b398c4c8f8e515341ac81d5c11e1b88348a2cff349306d7299c715065f1530cb"} Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.733924 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-kkks2" event={"ID":"92a07292-975e-492a-97da-f713cfb5dbce","Type":"ContainerDied","Data":"1d9f0e18f3c3062056a3554266354abe51a2816235d434999066da7714fd2efd"} Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.734017 4808 scope.go:117] "RemoveContainer" containerID="b398c4c8f8e515341ac81d5c11e1b88348a2cff349306d7299c715065f1530cb" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.734730 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.734951 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="2bf755e9-580a-497d-bb69-8f886fcb8dc4" containerName="probe" containerID="cri-o://20d11cfaabe7fd0fee6156d616541b3adb397d5a41ee987463fd71404b28fc08" gracePeriod=30 Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.770332 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-config\") pod \"92a07292-975e-492a-97da-f713cfb5dbce\" (UID: \"92a07292-975e-492a-97da-f713cfb5dbce\") " Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.770365 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-ovsdbserver-nb\") pod \"92a07292-975e-492a-97da-f713cfb5dbce\" (UID: \"92a07292-975e-492a-97da-f713cfb5dbce\") " Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.770484 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-dns-svc\") pod \"92a07292-975e-492a-97da-f713cfb5dbce\" (UID: \"92a07292-975e-492a-97da-f713cfb5dbce\") " Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.770545 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-ovsdbserver-sb\") pod \"92a07292-975e-492a-97da-f713cfb5dbce\" (UID: \"92a07292-975e-492a-97da-f713cfb5dbce\") " Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.770571 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mw2j\" (UniqueName: \"kubernetes.io/projected/92a07292-975e-492a-97da-f713cfb5dbce-kube-api-access-4mw2j\") pod \"92a07292-975e-492a-97da-f713cfb5dbce\" (UID: \"92a07292-975e-492a-97da-f713cfb5dbce\") " Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.785792 4808 scope.go:117] "RemoveContainer" containerID="eec0e4699b3d8afef7c91bba2e79f85f1a0d2dd638a7df19ddbf3a42b2d33619" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.789182 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92a07292-975e-492a-97da-f713cfb5dbce-kube-api-access-4mw2j" (OuterVolumeSpecName: "kube-api-access-4mw2j") pod "92a07292-975e-492a-97da-f713cfb5dbce" (UID: "92a07292-975e-492a-97da-f713cfb5dbce"). InnerVolumeSpecName "kube-api-access-4mw2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.823260 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-config" (OuterVolumeSpecName: "config") pod "92a07292-975e-492a-97da-f713cfb5dbce" (UID: "92a07292-975e-492a-97da-f713cfb5dbce"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.835492 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "92a07292-975e-492a-97da-f713cfb5dbce" (UID: "92a07292-975e-492a-97da-f713cfb5dbce"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.844597 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "92a07292-975e-492a-97da-f713cfb5dbce" (UID: "92a07292-975e-492a-97da-f713cfb5dbce"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.866032 4808 scope.go:117] "RemoveContainer" containerID="b398c4c8f8e515341ac81d5c11e1b88348a2cff349306d7299c715065f1530cb" Oct 02 16:58:37 crc kubenswrapper[4808]: E1002 16:58:37.866471 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b398c4c8f8e515341ac81d5c11e1b88348a2cff349306d7299c715065f1530cb\": container with ID starting with b398c4c8f8e515341ac81d5c11e1b88348a2cff349306d7299c715065f1530cb not found: ID does not exist" containerID="b398c4c8f8e515341ac81d5c11e1b88348a2cff349306d7299c715065f1530cb" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.866503 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b398c4c8f8e515341ac81d5c11e1b88348a2cff349306d7299c715065f1530cb"} err="failed to get container status \"b398c4c8f8e515341ac81d5c11e1b88348a2cff349306d7299c715065f1530cb\": rpc error: code = NotFound desc = could not find container \"b398c4c8f8e515341ac81d5c11e1b88348a2cff349306d7299c715065f1530cb\": container with ID starting with b398c4c8f8e515341ac81d5c11e1b88348a2cff349306d7299c715065f1530cb not found: ID does not exist" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.866522 4808 scope.go:117] "RemoveContainer" containerID="eec0e4699b3d8afef7c91bba2e79f85f1a0d2dd638a7df19ddbf3a42b2d33619" Oct 02 16:58:37 crc kubenswrapper[4808]: E1002 16:58:37.866814 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eec0e4699b3d8afef7c91bba2e79f85f1a0d2dd638a7df19ddbf3a42b2d33619\": container with ID starting with eec0e4699b3d8afef7c91bba2e79f85f1a0d2dd638a7df19ddbf3a42b2d33619 not found: ID does not exist" containerID="eec0e4699b3d8afef7c91bba2e79f85f1a0d2dd638a7df19ddbf3a42b2d33619" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.866872 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eec0e4699b3d8afef7c91bba2e79f85f1a0d2dd638a7df19ddbf3a42b2d33619"} err="failed to get container status \"eec0e4699b3d8afef7c91bba2e79f85f1a0d2dd638a7df19ddbf3a42b2d33619\": rpc error: code = NotFound desc = could not find container \"eec0e4699b3d8afef7c91bba2e79f85f1a0d2dd638a7df19ddbf3a42b2d33619\": container with ID starting with eec0e4699b3d8afef7c91bba2e79f85f1a0d2dd638a7df19ddbf3a42b2d33619 not found: ID does not exist" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.876066 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.876106 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mw2j\" (UniqueName: \"kubernetes.io/projected/92a07292-975e-492a-97da-f713cfb5dbce-kube-api-access-4mw2j\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.876135 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.876145 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.882990 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "92a07292-975e-492a-97da-f713cfb5dbce" (UID: "92a07292-975e-492a-97da-f713cfb5dbce"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:58:37 crc kubenswrapper[4808]: I1002 16:58:37.977956 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92a07292-975e-492a-97da-f713cfb5dbce-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:38 crc kubenswrapper[4808]: I1002 16:58:38.068572 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-kkks2"] Oct 02 16:58:38 crc kubenswrapper[4808]: I1002 16:58:38.076539 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-kkks2"] Oct 02 16:58:38 crc kubenswrapper[4808]: I1002 16:58:38.741378 4808 generic.go:334] "Generic (PLEG): container finished" podID="2bf755e9-580a-497d-bb69-8f886fcb8dc4" containerID="20d11cfaabe7fd0fee6156d616541b3adb397d5a41ee987463fd71404b28fc08" exitCode=0 Oct 02 16:58:38 crc kubenswrapper[4808]: I1002 16:58:38.741448 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2bf755e9-580a-497d-bb69-8f886fcb8dc4","Type":"ContainerDied","Data":"20d11cfaabe7fd0fee6156d616541b3adb397d5a41ee987463fd71404b28fc08"} Oct 02 16:58:39 crc kubenswrapper[4808]: I1002 16:58:39.407949 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92a07292-975e-492a-97da-f713cfb5dbce" path="/var/lib/kubelet/pods/92a07292-975e-492a-97da-f713cfb5dbce/volumes" Oct 02 16:58:41 crc kubenswrapper[4808]: I1002 16:58:41.778580 4808 generic.go:334] "Generic (PLEG): container finished" podID="2bf755e9-580a-497d-bb69-8f886fcb8dc4" containerID="d9a695a9740bee4636c6be6d2b8147c5bc18a38a613d89d626afbf205a7830c3" exitCode=0 Oct 02 16:58:41 crc kubenswrapper[4808]: I1002 16:58:41.779114 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2bf755e9-580a-497d-bb69-8f886fcb8dc4","Type":"ContainerDied","Data":"d9a695a9740bee4636c6be6d2b8147c5bc18a38a613d89d626afbf205a7830c3"} Oct 02 16:58:41 crc kubenswrapper[4808]: I1002 16:58:41.944448 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.065175 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpcrv\" (UniqueName: \"kubernetes.io/projected/2bf755e9-580a-497d-bb69-8f886fcb8dc4-kube-api-access-xpcrv\") pod \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.065299 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-scripts\") pod \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.065327 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-config-data\") pod \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.065358 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-config-data-custom\") pod \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.065455 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-combined-ca-bundle\") pod \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.065512 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2bf755e9-580a-497d-bb69-8f886fcb8dc4-etc-machine-id\") pod \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\" (UID: \"2bf755e9-580a-497d-bb69-8f886fcb8dc4\") " Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.065670 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2bf755e9-580a-497d-bb69-8f886fcb8dc4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2bf755e9-580a-497d-bb69-8f886fcb8dc4" (UID: "2bf755e9-580a-497d-bb69-8f886fcb8dc4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.066052 4808 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2bf755e9-580a-497d-bb69-8f886fcb8dc4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.087664 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-scripts" (OuterVolumeSpecName: "scripts") pod "2bf755e9-580a-497d-bb69-8f886fcb8dc4" (UID: "2bf755e9-580a-497d-bb69-8f886fcb8dc4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.089635 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2bf755e9-580a-497d-bb69-8f886fcb8dc4" (UID: "2bf755e9-580a-497d-bb69-8f886fcb8dc4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.095127 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bf755e9-580a-497d-bb69-8f886fcb8dc4-kube-api-access-xpcrv" (OuterVolumeSpecName: "kube-api-access-xpcrv") pod "2bf755e9-580a-497d-bb69-8f886fcb8dc4" (UID: "2bf755e9-580a-497d-bb69-8f886fcb8dc4"). InnerVolumeSpecName "kube-api-access-xpcrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.139108 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2bf755e9-580a-497d-bb69-8f886fcb8dc4" (UID: "2bf755e9-580a-497d-bb69-8f886fcb8dc4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.167658 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpcrv\" (UniqueName: \"kubernetes.io/projected/2bf755e9-580a-497d-bb69-8f886fcb8dc4-kube-api-access-xpcrv\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.167688 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.167700 4808 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.167710 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.168405 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-config-data" (OuterVolumeSpecName: "config-data") pod "2bf755e9-580a-497d-bb69-8f886fcb8dc4" (UID: "2bf755e9-580a-497d-bb69-8f886fcb8dc4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.269058 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bf755e9-580a-497d-bb69-8f886fcb8dc4-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.788197 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2bf755e9-580a-497d-bb69-8f886fcb8dc4","Type":"ContainerDied","Data":"95913a43205d00578535fc3ca12973f45a2f8733ef383aa8ca295486366889a4"} Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.788534 4808 scope.go:117] "RemoveContainer" containerID="20d11cfaabe7fd0fee6156d616541b3adb397d5a41ee987463fd71404b28fc08" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.788652 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.833157 4808 scope.go:117] "RemoveContainer" containerID="d9a695a9740bee4636c6be6d2b8147c5bc18a38a613d89d626afbf205a7830c3" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.855787 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.863203 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.876744 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 16:58:42 crc kubenswrapper[4808]: E1002 16:58:42.877154 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bf755e9-580a-497d-bb69-8f886fcb8dc4" containerName="cinder-scheduler" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.877173 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bf755e9-580a-497d-bb69-8f886fcb8dc4" containerName="cinder-scheduler" Oct 02 16:58:42 crc kubenswrapper[4808]: E1002 16:58:42.877186 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a07292-975e-492a-97da-f713cfb5dbce" containerName="init" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.877191 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a07292-975e-492a-97da-f713cfb5dbce" containerName="init" Oct 02 16:58:42 crc kubenswrapper[4808]: E1002 16:58:42.877208 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bf755e9-580a-497d-bb69-8f886fcb8dc4" containerName="probe" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.877214 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bf755e9-580a-497d-bb69-8f886fcb8dc4" containerName="probe" Oct 02 16:58:42 crc kubenswrapper[4808]: E1002 16:58:42.877226 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a07292-975e-492a-97da-f713cfb5dbce" containerName="dnsmasq-dns" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.877244 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a07292-975e-492a-97da-f713cfb5dbce" containerName="dnsmasq-dns" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.877404 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bf755e9-580a-497d-bb69-8f886fcb8dc4" containerName="cinder-scheduler" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.877421 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bf755e9-580a-497d-bb69-8f886fcb8dc4" containerName="probe" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.877434 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="92a07292-975e-492a-97da-f713cfb5dbce" containerName="dnsmasq-dns" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.878394 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.878767 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.882578 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.981220 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a608d71c-370e-46ca-bf97-1752628865ad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a608d71c-370e-46ca-bf97-1752628865ad\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.981318 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a608d71c-370e-46ca-bf97-1752628865ad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a608d71c-370e-46ca-bf97-1752628865ad\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.981354 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a608d71c-370e-46ca-bf97-1752628865ad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a608d71c-370e-46ca-bf97-1752628865ad\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.981401 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a608d71c-370e-46ca-bf97-1752628865ad-config-data\") pod \"cinder-scheduler-0\" (UID: \"a608d71c-370e-46ca-bf97-1752628865ad\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.981425 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a608d71c-370e-46ca-bf97-1752628865ad-scripts\") pod \"cinder-scheduler-0\" (UID: \"a608d71c-370e-46ca-bf97-1752628865ad\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:42 crc kubenswrapper[4808]: I1002 16:58:42.981472 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5brpj\" (UniqueName: \"kubernetes.io/projected/a608d71c-370e-46ca-bf97-1752628865ad-kube-api-access-5brpj\") pod \"cinder-scheduler-0\" (UID: \"a608d71c-370e-46ca-bf97-1752628865ad\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:43 crc kubenswrapper[4808]: I1002 16:58:43.082977 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a608d71c-370e-46ca-bf97-1752628865ad-config-data\") pod \"cinder-scheduler-0\" (UID: \"a608d71c-370e-46ca-bf97-1752628865ad\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:43 crc kubenswrapper[4808]: I1002 16:58:43.083028 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a608d71c-370e-46ca-bf97-1752628865ad-scripts\") pod \"cinder-scheduler-0\" (UID: \"a608d71c-370e-46ca-bf97-1752628865ad\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:43 crc kubenswrapper[4808]: I1002 16:58:43.083090 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5brpj\" (UniqueName: \"kubernetes.io/projected/a608d71c-370e-46ca-bf97-1752628865ad-kube-api-access-5brpj\") pod \"cinder-scheduler-0\" (UID: \"a608d71c-370e-46ca-bf97-1752628865ad\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:43 crc kubenswrapper[4808]: I1002 16:58:43.083135 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a608d71c-370e-46ca-bf97-1752628865ad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a608d71c-370e-46ca-bf97-1752628865ad\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:43 crc kubenswrapper[4808]: I1002 16:58:43.083168 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a608d71c-370e-46ca-bf97-1752628865ad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a608d71c-370e-46ca-bf97-1752628865ad\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:43 crc kubenswrapper[4808]: I1002 16:58:43.083198 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a608d71c-370e-46ca-bf97-1752628865ad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a608d71c-370e-46ca-bf97-1752628865ad\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:43 crc kubenswrapper[4808]: I1002 16:58:43.083293 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a608d71c-370e-46ca-bf97-1752628865ad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a608d71c-370e-46ca-bf97-1752628865ad\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:43 crc kubenswrapper[4808]: I1002 16:58:43.091956 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a608d71c-370e-46ca-bf97-1752628865ad-scripts\") pod \"cinder-scheduler-0\" (UID: \"a608d71c-370e-46ca-bf97-1752628865ad\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:43 crc kubenswrapper[4808]: I1002 16:58:43.092226 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a608d71c-370e-46ca-bf97-1752628865ad-config-data\") pod \"cinder-scheduler-0\" (UID: \"a608d71c-370e-46ca-bf97-1752628865ad\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:43 crc kubenswrapper[4808]: I1002 16:58:43.097760 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a608d71c-370e-46ca-bf97-1752628865ad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a608d71c-370e-46ca-bf97-1752628865ad\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:43 crc kubenswrapper[4808]: I1002 16:58:43.100721 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a608d71c-370e-46ca-bf97-1752628865ad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a608d71c-370e-46ca-bf97-1752628865ad\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:43 crc kubenswrapper[4808]: I1002 16:58:43.104880 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5brpj\" (UniqueName: \"kubernetes.io/projected/a608d71c-370e-46ca-bf97-1752628865ad-kube-api-access-5brpj\") pod \"cinder-scheduler-0\" (UID: \"a608d71c-370e-46ca-bf97-1752628865ad\") " pod="openstack/cinder-scheduler-0" Oct 02 16:58:43 crc kubenswrapper[4808]: I1002 16:58:43.213110 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 16:58:43 crc kubenswrapper[4808]: I1002 16:58:43.429557 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bf755e9-580a-497d-bb69-8f886fcb8dc4" path="/var/lib/kubelet/pods/2bf755e9-580a-497d-bb69-8f886fcb8dc4/volumes" Oct 02 16:58:43 crc kubenswrapper[4808]: I1002 16:58:43.739139 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:43 crc kubenswrapper[4808]: I1002 16:58:43.841777 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 16:58:43 crc kubenswrapper[4808]: I1002 16:58:43.902822 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5684d76658-g9f5f" Oct 02 16:58:43 crc kubenswrapper[4808]: I1002 16:58:43.967261 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-798c55655d-t754s"] Oct 02 16:58:43 crc kubenswrapper[4808]: I1002 16:58:43.967854 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-798c55655d-t754s" podUID="28edd780-f606-40e4-a9a3-babce79c3f0f" containerName="barbican-api-log" containerID="cri-o://ddb6929dc87352bd47c3e477cd0f5d8ef42ac3c2f630251e5c5634ec853b998e" gracePeriod=30 Oct 02 16:58:43 crc kubenswrapper[4808]: I1002 16:58:43.968427 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-798c55655d-t754s" podUID="28edd780-f606-40e4-a9a3-babce79c3f0f" containerName="barbican-api" containerID="cri-o://d9203c1a421d591e83f7a8078f645d2b588113534092e579d107f92b6d697304" gracePeriod=30 Oct 02 16:58:44 crc kubenswrapper[4808]: I1002 16:58:44.510012 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-577d5bdf64-z4hwp" Oct 02 16:58:44 crc kubenswrapper[4808]: I1002 16:58:44.829614 4808 generic.go:334] "Generic (PLEG): container finished" podID="28edd780-f606-40e4-a9a3-babce79c3f0f" containerID="ddb6929dc87352bd47c3e477cd0f5d8ef42ac3c2f630251e5c5634ec853b998e" exitCode=143 Oct 02 16:58:44 crc kubenswrapper[4808]: I1002 16:58:44.829695 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-798c55655d-t754s" event={"ID":"28edd780-f606-40e4-a9a3-babce79c3f0f","Type":"ContainerDied","Data":"ddb6929dc87352bd47c3e477cd0f5d8ef42ac3c2f630251e5c5634ec853b998e"} Oct 02 16:58:44 crc kubenswrapper[4808]: I1002 16:58:44.834340 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a608d71c-370e-46ca-bf97-1752628865ad","Type":"ContainerStarted","Data":"93796f003b4e4e824973167b211c17d5e2e57e52f20976239d3eef29eb60491f"} Oct 02 16:58:44 crc kubenswrapper[4808]: I1002 16:58:44.834386 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a608d71c-370e-46ca-bf97-1752628865ad","Type":"ContainerStarted","Data":"9c8bbad07501264c0fe2792388771dc2524b917ab128cd36c5b3ea75206da54e"} Oct 02 16:58:44 crc kubenswrapper[4808]: I1002 16:58:44.836715 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:44 crc kubenswrapper[4808]: I1002 16:58:44.911721 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 02 16:58:44 crc kubenswrapper[4808]: I1002 16:58:44.912813 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 16:58:44 crc kubenswrapper[4808]: I1002 16:58:44.918791 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 02 16:58:44 crc kubenswrapper[4808]: I1002 16:58:44.919414 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 02 16:58:44 crc kubenswrapper[4808]: I1002 16:58:44.920594 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-h6f57" Oct 02 16:58:44 crc kubenswrapper[4808]: I1002 16:58:44.923214 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 16:58:45 crc kubenswrapper[4808]: I1002 16:58:45.046277 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79b74b54-a9e1-4a7c-841a-1b4d4588b6bf-openstack-config-secret\") pod \"openstackclient\" (UID: \"79b74b54-a9e1-4a7c-841a-1b4d4588b6bf\") " pod="openstack/openstackclient" Oct 02 16:58:45 crc kubenswrapper[4808]: I1002 16:58:45.046593 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b74b54-a9e1-4a7c-841a-1b4d4588b6bf-combined-ca-bundle\") pod \"openstackclient\" (UID: \"79b74b54-a9e1-4a7c-841a-1b4d4588b6bf\") " pod="openstack/openstackclient" Oct 02 16:58:45 crc kubenswrapper[4808]: I1002 16:58:45.046649 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp9lz\" (UniqueName: \"kubernetes.io/projected/79b74b54-a9e1-4a7c-841a-1b4d4588b6bf-kube-api-access-sp9lz\") pod \"openstackclient\" (UID: \"79b74b54-a9e1-4a7c-841a-1b4d4588b6bf\") " pod="openstack/openstackclient" Oct 02 16:58:45 crc kubenswrapper[4808]: I1002 16:58:45.046683 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/79b74b54-a9e1-4a7c-841a-1b4d4588b6bf-openstack-config\") pod \"openstackclient\" (UID: \"79b74b54-a9e1-4a7c-841a-1b4d4588b6bf\") " pod="openstack/openstackclient" Oct 02 16:58:45 crc kubenswrapper[4808]: I1002 16:58:45.051834 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7dbcfb7656-7hqb7" Oct 02 16:58:45 crc kubenswrapper[4808]: I1002 16:58:45.154152 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79b74b54-a9e1-4a7c-841a-1b4d4588b6bf-openstack-config-secret\") pod \"openstackclient\" (UID: \"79b74b54-a9e1-4a7c-841a-1b4d4588b6bf\") " pod="openstack/openstackclient" Oct 02 16:58:45 crc kubenswrapper[4808]: I1002 16:58:45.154252 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b74b54-a9e1-4a7c-841a-1b4d4588b6bf-combined-ca-bundle\") pod \"openstackclient\" (UID: \"79b74b54-a9e1-4a7c-841a-1b4d4588b6bf\") " pod="openstack/openstackclient" Oct 02 16:58:45 crc kubenswrapper[4808]: I1002 16:58:45.154454 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp9lz\" (UniqueName: \"kubernetes.io/projected/79b74b54-a9e1-4a7c-841a-1b4d4588b6bf-kube-api-access-sp9lz\") pod \"openstackclient\" (UID: \"79b74b54-a9e1-4a7c-841a-1b4d4588b6bf\") " pod="openstack/openstackclient" Oct 02 16:58:45 crc kubenswrapper[4808]: I1002 16:58:45.154546 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/79b74b54-a9e1-4a7c-841a-1b4d4588b6bf-openstack-config\") pod \"openstackclient\" (UID: \"79b74b54-a9e1-4a7c-841a-1b4d4588b6bf\") " pod="openstack/openstackclient" Oct 02 16:58:45 crc kubenswrapper[4808]: I1002 16:58:45.156431 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/79b74b54-a9e1-4a7c-841a-1b4d4588b6bf-openstack-config\") pod \"openstackclient\" (UID: \"79b74b54-a9e1-4a7c-841a-1b4d4588b6bf\") " pod="openstack/openstackclient" Oct 02 16:58:45 crc kubenswrapper[4808]: I1002 16:58:45.159652 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79b74b54-a9e1-4a7c-841a-1b4d4588b6bf-openstack-config-secret\") pod \"openstackclient\" (UID: \"79b74b54-a9e1-4a7c-841a-1b4d4588b6bf\") " pod="openstack/openstackclient" Oct 02 16:58:45 crc kubenswrapper[4808]: I1002 16:58:45.163641 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b74b54-a9e1-4a7c-841a-1b4d4588b6bf-combined-ca-bundle\") pod \"openstackclient\" (UID: \"79b74b54-a9e1-4a7c-841a-1b4d4588b6bf\") " pod="openstack/openstackclient" Oct 02 16:58:45 crc kubenswrapper[4808]: I1002 16:58:45.187799 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp9lz\" (UniqueName: \"kubernetes.io/projected/79b74b54-a9e1-4a7c-841a-1b4d4588b6bf-kube-api-access-sp9lz\") pod \"openstackclient\" (UID: \"79b74b54-a9e1-4a7c-841a-1b4d4588b6bf\") " pod="openstack/openstackclient" Oct 02 16:58:45 crc kubenswrapper[4808]: I1002 16:58:45.248041 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 16:58:45 crc kubenswrapper[4808]: I1002 16:58:45.773535 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 16:58:45 crc kubenswrapper[4808]: W1002 16:58:45.781730 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79b74b54_a9e1_4a7c_841a_1b4d4588b6bf.slice/crio-a6c42ab1c7c26086b5b2728a8e3822048aa84ecb24d009b60bb2a459feb89492 WatchSource:0}: Error finding container a6c42ab1c7c26086b5b2728a8e3822048aa84ecb24d009b60bb2a459feb89492: Status 404 returned error can't find the container with id a6c42ab1c7c26086b5b2728a8e3822048aa84ecb24d009b60bb2a459feb89492 Oct 02 16:58:45 crc kubenswrapper[4808]: I1002 16:58:45.848338 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"79b74b54-a9e1-4a7c-841a-1b4d4588b6bf","Type":"ContainerStarted","Data":"a6c42ab1c7c26086b5b2728a8e3822048aa84ecb24d009b60bb2a459feb89492"} Oct 02 16:58:45 crc kubenswrapper[4808]: I1002 16:58:45.851715 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a608d71c-370e-46ca-bf97-1752628865ad","Type":"ContainerStarted","Data":"9b03b7fff844ecfe13375e112db6da98f708d4059c189410f88ba2f1f5c2b820"} Oct 02 16:58:45 crc kubenswrapper[4808]: I1002 16:58:45.871850 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.871829236 podStartE2EDuration="3.871829236s" podCreationTimestamp="2025-10-02 16:58:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:58:45.865369191 +0000 UTC m=+1093.190898211" watchObservedRunningTime="2025-10-02 16:58:45.871829236 +0000 UTC m=+1093.197358246" Oct 02 16:58:46 crc kubenswrapper[4808]: I1002 16:58:46.457044 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.144839 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-798c55655d-t754s" podUID="28edd780-f606-40e4-a9a3-babce79c3f0f" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.149:9311/healthcheck\": read tcp 10.217.0.2:46216->10.217.0.149:9311: read: connection reset by peer" Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.144905 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-798c55655d-t754s" podUID="28edd780-f606-40e4-a9a3-babce79c3f0f" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.149:9311/healthcheck\": read tcp 10.217.0.2:46212->10.217.0.149:9311: read: connection reset by peer" Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.656221 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.718277 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/28edd780-f606-40e4-a9a3-babce79c3f0f-config-data-custom\") pod \"28edd780-f606-40e4-a9a3-babce79c3f0f\" (UID: \"28edd780-f606-40e4-a9a3-babce79c3f0f\") " Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.718439 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28edd780-f606-40e4-a9a3-babce79c3f0f-logs\") pod \"28edd780-f606-40e4-a9a3-babce79c3f0f\" (UID: \"28edd780-f606-40e4-a9a3-babce79c3f0f\") " Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.718469 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28edd780-f606-40e4-a9a3-babce79c3f0f-config-data\") pod \"28edd780-f606-40e4-a9a3-babce79c3f0f\" (UID: \"28edd780-f606-40e4-a9a3-babce79c3f0f\") " Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.718502 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28edd780-f606-40e4-a9a3-babce79c3f0f-combined-ca-bundle\") pod \"28edd780-f606-40e4-a9a3-babce79c3f0f\" (UID: \"28edd780-f606-40e4-a9a3-babce79c3f0f\") " Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.718526 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mf6v\" (UniqueName: \"kubernetes.io/projected/28edd780-f606-40e4-a9a3-babce79c3f0f-kube-api-access-8mf6v\") pod \"28edd780-f606-40e4-a9a3-babce79c3f0f\" (UID: \"28edd780-f606-40e4-a9a3-babce79c3f0f\") " Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.719268 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28edd780-f606-40e4-a9a3-babce79c3f0f-logs" (OuterVolumeSpecName: "logs") pod "28edd780-f606-40e4-a9a3-babce79c3f0f" (UID: "28edd780-f606-40e4-a9a3-babce79c3f0f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.738498 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28edd780-f606-40e4-a9a3-babce79c3f0f-kube-api-access-8mf6v" (OuterVolumeSpecName: "kube-api-access-8mf6v") pod "28edd780-f606-40e4-a9a3-babce79c3f0f" (UID: "28edd780-f606-40e4-a9a3-babce79c3f0f"). InnerVolumeSpecName "kube-api-access-8mf6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.739868 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28edd780-f606-40e4-a9a3-babce79c3f0f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "28edd780-f606-40e4-a9a3-babce79c3f0f" (UID: "28edd780-f606-40e4-a9a3-babce79c3f0f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.746186 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28edd780-f606-40e4-a9a3-babce79c3f0f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "28edd780-f606-40e4-a9a3-babce79c3f0f" (UID: "28edd780-f606-40e4-a9a3-babce79c3f0f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.783184 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28edd780-f606-40e4-a9a3-babce79c3f0f-config-data" (OuterVolumeSpecName: "config-data") pod "28edd780-f606-40e4-a9a3-babce79c3f0f" (UID: "28edd780-f606-40e4-a9a3-babce79c3f0f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.819931 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28edd780-f606-40e4-a9a3-babce79c3f0f-logs\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.820000 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28edd780-f606-40e4-a9a3-babce79c3f0f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.820014 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28edd780-f606-40e4-a9a3-babce79c3f0f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.820025 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mf6v\" (UniqueName: \"kubernetes.io/projected/28edd780-f606-40e4-a9a3-babce79c3f0f-kube-api-access-8mf6v\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.820034 4808 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/28edd780-f606-40e4-a9a3-babce79c3f0f-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.872538 4808 generic.go:334] "Generic (PLEG): container finished" podID="28edd780-f606-40e4-a9a3-babce79c3f0f" containerID="d9203c1a421d591e83f7a8078f645d2b588113534092e579d107f92b6d697304" exitCode=0 Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.872591 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-798c55655d-t754s" event={"ID":"28edd780-f606-40e4-a9a3-babce79c3f0f","Type":"ContainerDied","Data":"d9203c1a421d591e83f7a8078f645d2b588113534092e579d107f92b6d697304"} Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.872620 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-798c55655d-t754s" event={"ID":"28edd780-f606-40e4-a9a3-babce79c3f0f","Type":"ContainerDied","Data":"8e2fa9afe1086d68dafc20560b71896db7692c239f11b186ff876d372c2bcf8b"} Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.872651 4808 scope.go:117] "RemoveContainer" containerID="d9203c1a421d591e83f7a8078f645d2b588113534092e579d107f92b6d697304" Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.872813 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-798c55655d-t754s" Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.907189 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-798c55655d-t754s"] Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.913411 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-798c55655d-t754s"] Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.918770 4808 scope.go:117] "RemoveContainer" containerID="ddb6929dc87352bd47c3e477cd0f5d8ef42ac3c2f630251e5c5634ec853b998e" Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.935577 4808 scope.go:117] "RemoveContainer" containerID="d9203c1a421d591e83f7a8078f645d2b588113534092e579d107f92b6d697304" Oct 02 16:58:47 crc kubenswrapper[4808]: E1002 16:58:47.936029 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9203c1a421d591e83f7a8078f645d2b588113534092e579d107f92b6d697304\": container with ID starting with d9203c1a421d591e83f7a8078f645d2b588113534092e579d107f92b6d697304 not found: ID does not exist" containerID="d9203c1a421d591e83f7a8078f645d2b588113534092e579d107f92b6d697304" Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.936059 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9203c1a421d591e83f7a8078f645d2b588113534092e579d107f92b6d697304"} err="failed to get container status \"d9203c1a421d591e83f7a8078f645d2b588113534092e579d107f92b6d697304\": rpc error: code = NotFound desc = could not find container \"d9203c1a421d591e83f7a8078f645d2b588113534092e579d107f92b6d697304\": container with ID starting with d9203c1a421d591e83f7a8078f645d2b588113534092e579d107f92b6d697304 not found: ID does not exist" Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.936081 4808 scope.go:117] "RemoveContainer" containerID="ddb6929dc87352bd47c3e477cd0f5d8ef42ac3c2f630251e5c5634ec853b998e" Oct 02 16:58:47 crc kubenswrapper[4808]: E1002 16:58:47.936403 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddb6929dc87352bd47c3e477cd0f5d8ef42ac3c2f630251e5c5634ec853b998e\": container with ID starting with ddb6929dc87352bd47c3e477cd0f5d8ef42ac3c2f630251e5c5634ec853b998e not found: ID does not exist" containerID="ddb6929dc87352bd47c3e477cd0f5d8ef42ac3c2f630251e5c5634ec853b998e" Oct 02 16:58:47 crc kubenswrapper[4808]: I1002 16:58:47.936429 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddb6929dc87352bd47c3e477cd0f5d8ef42ac3c2f630251e5c5634ec853b998e"} err="failed to get container status \"ddb6929dc87352bd47c3e477cd0f5d8ef42ac3c2f630251e5c5634ec853b998e\": rpc error: code = NotFound desc = could not find container \"ddb6929dc87352bd47c3e477cd0f5d8ef42ac3c2f630251e5c5634ec853b998e\": container with ID starting with ddb6929dc87352bd47c3e477cd0f5d8ef42ac3c2f630251e5c5634ec853b998e not found: ID does not exist" Oct 02 16:58:48 crc kubenswrapper[4808]: I1002 16:58:48.213205 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 16:58:49 crc kubenswrapper[4808]: I1002 16:58:49.406817 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28edd780-f606-40e4-a9a3-babce79c3f0f" path="/var/lib/kubelet/pods/28edd780-f606-40e4-a9a3-babce79c3f0f/volumes" Oct 02 16:58:50 crc kubenswrapper[4808]: I1002 16:58:50.641106 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:52 crc kubenswrapper[4808]: I1002 16:58:52.255548 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:58:52 crc kubenswrapper[4808]: I1002 16:58:52.256144 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c16fcba0-e224-4b8f-a220-71184caeede4" containerName="ceilometer-central-agent" containerID="cri-o://2587f9f7ff50a6b9964bb34e849273fe79595b5a4f5831e89e480a2abe53d4ea" gracePeriod=30 Oct 02 16:58:52 crc kubenswrapper[4808]: I1002 16:58:52.256645 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c16fcba0-e224-4b8f-a220-71184caeede4" containerName="proxy-httpd" containerID="cri-o://c488da3f08e05efc006ef52801722d928ad4f2a86704dfb0153b85eafaa838f6" gracePeriod=30 Oct 02 16:58:52 crc kubenswrapper[4808]: I1002 16:58:52.256699 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c16fcba0-e224-4b8f-a220-71184caeede4" containerName="sg-core" containerID="cri-o://85f033100798bc2c00717d786e992f63420eb42ea2853f1627165cf4d08e2a06" gracePeriod=30 Oct 02 16:58:52 crc kubenswrapper[4808]: I1002 16:58:52.256740 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c16fcba0-e224-4b8f-a220-71184caeede4" containerName="ceilometer-notification-agent" containerID="cri-o://438653f745c160c504df307e4d593d42e716a38f68e29354622a4c8a163a689c" gracePeriod=30 Oct 02 16:58:52 crc kubenswrapper[4808]: I1002 16:58:52.266291 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="c16fcba0-e224-4b8f-a220-71184caeede4" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.145:3000/\": EOF" Oct 02 16:58:52 crc kubenswrapper[4808]: I1002 16:58:52.843451 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-78b67777b5-h245x" Oct 02 16:58:52 crc kubenswrapper[4808]: I1002 16:58:52.946042 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c16fcba0-e224-4b8f-a220-71184caeede4","Type":"ContainerDied","Data":"c488da3f08e05efc006ef52801722d928ad4f2a86704dfb0153b85eafaa838f6"} Oct 02 16:58:52 crc kubenswrapper[4808]: I1002 16:58:52.946120 4808 generic.go:334] "Generic (PLEG): container finished" podID="c16fcba0-e224-4b8f-a220-71184caeede4" containerID="c488da3f08e05efc006ef52801722d928ad4f2a86704dfb0153b85eafaa838f6" exitCode=0 Oct 02 16:58:52 crc kubenswrapper[4808]: I1002 16:58:52.946143 4808 generic.go:334] "Generic (PLEG): container finished" podID="c16fcba0-e224-4b8f-a220-71184caeede4" containerID="85f033100798bc2c00717d786e992f63420eb42ea2853f1627165cf4d08e2a06" exitCode=2 Oct 02 16:58:52 crc kubenswrapper[4808]: I1002 16:58:52.946151 4808 generic.go:334] "Generic (PLEG): container finished" podID="c16fcba0-e224-4b8f-a220-71184caeede4" containerID="2587f9f7ff50a6b9964bb34e849273fe79595b5a4f5831e89e480a2abe53d4ea" exitCode=0 Oct 02 16:58:52 crc kubenswrapper[4808]: I1002 16:58:52.946165 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c16fcba0-e224-4b8f-a220-71184caeede4","Type":"ContainerDied","Data":"85f033100798bc2c00717d786e992f63420eb42ea2853f1627165cf4d08e2a06"} Oct 02 16:58:52 crc kubenswrapper[4808]: I1002 16:58:52.946434 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c16fcba0-e224-4b8f-a220-71184caeede4","Type":"ContainerDied","Data":"2587f9f7ff50a6b9964bb34e849273fe79595b5a4f5831e89e480a2abe53d4ea"} Oct 02 16:58:52 crc kubenswrapper[4808]: I1002 16:58:52.962862 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-59d95756b-rl9bd"] Oct 02 16:58:52 crc kubenswrapper[4808]: I1002 16:58:52.963109 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-59d95756b-rl9bd" podUID="688a2bb4-4a3a-4a13-87a8-fbf59024dd3d" containerName="neutron-api" containerID="cri-o://1472f9b375176e20444d88bf3d34fe527360069ebe368cd9ad0f391466df3b5b" gracePeriod=30 Oct 02 16:58:52 crc kubenswrapper[4808]: I1002 16:58:52.963607 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-59d95756b-rl9bd" podUID="688a2bb4-4a3a-4a13-87a8-fbf59024dd3d" containerName="neutron-httpd" containerID="cri-o://7da26f1db3010f3f7f174cddb5acd13403301a6470274ecef4756758ffdc9612" gracePeriod=30 Oct 02 16:58:53 crc kubenswrapper[4808]: I1002 16:58:53.484282 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 16:58:53 crc kubenswrapper[4808]: I1002 16:58:53.959553 4808 generic.go:334] "Generic (PLEG): container finished" podID="688a2bb4-4a3a-4a13-87a8-fbf59024dd3d" containerID="7da26f1db3010f3f7f174cddb5acd13403301a6470274ecef4756758ffdc9612" exitCode=0 Oct 02 16:58:53 crc kubenswrapper[4808]: I1002 16:58:53.959736 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59d95756b-rl9bd" event={"ID":"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d","Type":"ContainerDied","Data":"7da26f1db3010f3f7f174cddb5acd13403301a6470274ecef4756758ffdc9612"} Oct 02 16:58:54 crc kubenswrapper[4808]: I1002 16:58:54.881469 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="c16fcba0-e224-4b8f-a220-71184caeede4" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.145:3000/\": dial tcp 10.217.0.145:3000: connect: connection refused" Oct 02 16:58:54 crc kubenswrapper[4808]: I1002 16:58:54.971561 4808 generic.go:334] "Generic (PLEG): container finished" podID="c16fcba0-e224-4b8f-a220-71184caeede4" containerID="438653f745c160c504df307e4d593d42e716a38f68e29354622a4c8a163a689c" exitCode=0 Oct 02 16:58:54 crc kubenswrapper[4808]: I1002 16:58:54.971602 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c16fcba0-e224-4b8f-a220-71184caeede4","Type":"ContainerDied","Data":"438653f745c160c504df307e4d593d42e716a38f68e29354622a4c8a163a689c"} Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.579552 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.670707 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-scripts\") pod \"c16fcba0-e224-4b8f-a220-71184caeede4\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.670792 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-config-data\") pod \"c16fcba0-e224-4b8f-a220-71184caeede4\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.670823 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-sg-core-conf-yaml\") pod \"c16fcba0-e224-4b8f-a220-71184caeede4\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.671382 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c16fcba0-e224-4b8f-a220-71184caeede4-log-httpd\") pod \"c16fcba0-e224-4b8f-a220-71184caeede4\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.671445 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c16fcba0-e224-4b8f-a220-71184caeede4-run-httpd\") pod \"c16fcba0-e224-4b8f-a220-71184caeede4\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.671683 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-combined-ca-bundle\") pod \"c16fcba0-e224-4b8f-a220-71184caeede4\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.671757 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dp4w2\" (UniqueName: \"kubernetes.io/projected/c16fcba0-e224-4b8f-a220-71184caeede4-kube-api-access-dp4w2\") pod \"c16fcba0-e224-4b8f-a220-71184caeede4\" (UID: \"c16fcba0-e224-4b8f-a220-71184caeede4\") " Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.671862 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c16fcba0-e224-4b8f-a220-71184caeede4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c16fcba0-e224-4b8f-a220-71184caeede4" (UID: "c16fcba0-e224-4b8f-a220-71184caeede4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.671962 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c16fcba0-e224-4b8f-a220-71184caeede4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c16fcba0-e224-4b8f-a220-71184caeede4" (UID: "c16fcba0-e224-4b8f-a220-71184caeede4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.672592 4808 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c16fcba0-e224-4b8f-a220-71184caeede4-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.672616 4808 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c16fcba0-e224-4b8f-a220-71184caeede4-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.674655 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-scripts" (OuterVolumeSpecName: "scripts") pod "c16fcba0-e224-4b8f-a220-71184caeede4" (UID: "c16fcba0-e224-4b8f-a220-71184caeede4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.677404 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c16fcba0-e224-4b8f-a220-71184caeede4-kube-api-access-dp4w2" (OuterVolumeSpecName: "kube-api-access-dp4w2") pod "c16fcba0-e224-4b8f-a220-71184caeede4" (UID: "c16fcba0-e224-4b8f-a220-71184caeede4"). InnerVolumeSpecName "kube-api-access-dp4w2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.704467 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c16fcba0-e224-4b8f-a220-71184caeede4" (UID: "c16fcba0-e224-4b8f-a220-71184caeede4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.757662 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c16fcba0-e224-4b8f-a220-71184caeede4" (UID: "c16fcba0-e224-4b8f-a220-71184caeede4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.774123 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.774157 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dp4w2\" (UniqueName: \"kubernetes.io/projected/c16fcba0-e224-4b8f-a220-71184caeede4-kube-api-access-dp4w2\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.774168 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.774178 4808 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.774310 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-config-data" (OuterVolumeSpecName: "config-data") pod "c16fcba0-e224-4b8f-a220-71184caeede4" (UID: "c16fcba0-e224-4b8f-a220-71184caeede4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:56 crc kubenswrapper[4808]: I1002 16:58:56.876433 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16fcba0-e224-4b8f-a220-71184caeede4-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.002833 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"79b74b54-a9e1-4a7c-841a-1b4d4588b6bf","Type":"ContainerStarted","Data":"e77b7a4dca288b449d8572c4682d4863dee11e693a9ce483c14479f5a575e998"} Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.006506 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c16fcba0-e224-4b8f-a220-71184caeede4","Type":"ContainerDied","Data":"18124b83073c3a542a8d914cb41ef4cb6e5710a9c65c7970f3cdcdaec177f5ac"} Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.006537 4808 scope.go:117] "RemoveContainer" containerID="c488da3f08e05efc006ef52801722d928ad4f2a86704dfb0153b85eafaa838f6" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.006630 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.049074 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.461106122 podStartE2EDuration="13.049048281s" podCreationTimestamp="2025-10-02 16:58:44 +0000 UTC" firstStartedPulling="2025-10-02 16:58:45.783607701 +0000 UTC m=+1093.109136711" lastFinishedPulling="2025-10-02 16:58:56.37154987 +0000 UTC m=+1103.697078870" observedRunningTime="2025-10-02 16:58:57.026257022 +0000 UTC m=+1104.351786022" watchObservedRunningTime="2025-10-02 16:58:57.049048281 +0000 UTC m=+1104.374577301" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.054573 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.060747 4808 scope.go:117] "RemoveContainer" containerID="85f033100798bc2c00717d786e992f63420eb42ea2853f1627165cf4d08e2a06" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.062002 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.082975 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:58:57 crc kubenswrapper[4808]: E1002 16:58:57.083335 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c16fcba0-e224-4b8f-a220-71184caeede4" containerName="sg-core" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.083354 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c16fcba0-e224-4b8f-a220-71184caeede4" containerName="sg-core" Oct 02 16:58:57 crc kubenswrapper[4808]: E1002 16:58:57.083372 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28edd780-f606-40e4-a9a3-babce79c3f0f" containerName="barbican-api" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.083379 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="28edd780-f606-40e4-a9a3-babce79c3f0f" containerName="barbican-api" Oct 02 16:58:57 crc kubenswrapper[4808]: E1002 16:58:57.083391 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c16fcba0-e224-4b8f-a220-71184caeede4" containerName="ceilometer-central-agent" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.083397 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c16fcba0-e224-4b8f-a220-71184caeede4" containerName="ceilometer-central-agent" Oct 02 16:58:57 crc kubenswrapper[4808]: E1002 16:58:57.083412 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28edd780-f606-40e4-a9a3-babce79c3f0f" containerName="barbican-api-log" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.083418 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="28edd780-f606-40e4-a9a3-babce79c3f0f" containerName="barbican-api-log" Oct 02 16:58:57 crc kubenswrapper[4808]: E1002 16:58:57.083432 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c16fcba0-e224-4b8f-a220-71184caeede4" containerName="proxy-httpd" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.083439 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c16fcba0-e224-4b8f-a220-71184caeede4" containerName="proxy-httpd" Oct 02 16:58:57 crc kubenswrapper[4808]: E1002 16:58:57.083454 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c16fcba0-e224-4b8f-a220-71184caeede4" containerName="ceilometer-notification-agent" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.083460 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c16fcba0-e224-4b8f-a220-71184caeede4" containerName="ceilometer-notification-agent" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.083614 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="28edd780-f606-40e4-a9a3-babce79c3f0f" containerName="barbican-api" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.083627 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c16fcba0-e224-4b8f-a220-71184caeede4" containerName="proxy-httpd" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.083642 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="28edd780-f606-40e4-a9a3-babce79c3f0f" containerName="barbican-api-log" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.083656 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c16fcba0-e224-4b8f-a220-71184caeede4" containerName="ceilometer-central-agent" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.083665 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c16fcba0-e224-4b8f-a220-71184caeede4" containerName="ceilometer-notification-agent" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.083675 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c16fcba0-e224-4b8f-a220-71184caeede4" containerName="sg-core" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.085498 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.087704 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.092435 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.105379 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.127131 4808 scope.go:117] "RemoveContainer" containerID="438653f745c160c504df307e4d593d42e716a38f68e29354622a4c8a163a689c" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.153945 4808 scope.go:117] "RemoveContainer" containerID="2587f9f7ff50a6b9964bb34e849273fe79595b5a4f5831e89e480a2abe53d4ea" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.180215 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56fad39c-0f30-463b-ab11-939c07877fcb-log-httpd\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.180265 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-scripts\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.180569 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g44tx\" (UniqueName: \"kubernetes.io/projected/56fad39c-0f30-463b-ab11-939c07877fcb-kube-api-access-g44tx\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.180726 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-config-data\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.180943 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.181034 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56fad39c-0f30-463b-ab11-939c07877fcb-run-httpd\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.181143 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.283136 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.283211 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56fad39c-0f30-463b-ab11-939c07877fcb-log-httpd\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.283246 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-scripts\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.283290 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g44tx\" (UniqueName: \"kubernetes.io/projected/56fad39c-0f30-463b-ab11-939c07877fcb-kube-api-access-g44tx\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.283319 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-config-data\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.283384 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.283411 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56fad39c-0f30-463b-ab11-939c07877fcb-run-httpd\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.283871 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56fad39c-0f30-463b-ab11-939c07877fcb-run-httpd\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.284773 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56fad39c-0f30-463b-ab11-939c07877fcb-log-httpd\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.287953 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-config-data\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.287980 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.288797 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-scripts\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.288833 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.305563 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g44tx\" (UniqueName: \"kubernetes.io/projected/56fad39c-0f30-463b-ab11-939c07877fcb-kube-api-access-g44tx\") pod \"ceilometer-0\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.422086 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c16fcba0-e224-4b8f-a220-71184caeede4" path="/var/lib/kubelet/pods/c16fcba0-e224-4b8f-a220-71184caeede4/volumes" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.448871 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.576539 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.702500 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqw2p\" (UniqueName: \"kubernetes.io/projected/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-kube-api-access-vqw2p\") pod \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\" (UID: \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\") " Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.702631 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-combined-ca-bundle\") pod \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\" (UID: \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\") " Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.702686 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-httpd-config\") pod \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\" (UID: \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\") " Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.702746 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-ovndb-tls-certs\") pod \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\" (UID: \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\") " Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.702775 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-config\") pod \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\" (UID: \"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d\") " Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.710218 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-kube-api-access-vqw2p" (OuterVolumeSpecName: "kube-api-access-vqw2p") pod "688a2bb4-4a3a-4a13-87a8-fbf59024dd3d" (UID: "688a2bb4-4a3a-4a13-87a8-fbf59024dd3d"). InnerVolumeSpecName "kube-api-access-vqw2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.712741 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "688a2bb4-4a3a-4a13-87a8-fbf59024dd3d" (UID: "688a2bb4-4a3a-4a13-87a8-fbf59024dd3d"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.758540 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-config" (OuterVolumeSpecName: "config") pod "688a2bb4-4a3a-4a13-87a8-fbf59024dd3d" (UID: "688a2bb4-4a3a-4a13-87a8-fbf59024dd3d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.762025 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "688a2bb4-4a3a-4a13-87a8-fbf59024dd3d" (UID: "688a2bb4-4a3a-4a13-87a8-fbf59024dd3d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.782891 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "688a2bb4-4a3a-4a13-87a8-fbf59024dd3d" (UID: "688a2bb4-4a3a-4a13-87a8-fbf59024dd3d"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.805253 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.805289 4808 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.805297 4808 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.805306 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.805316 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqw2p\" (UniqueName: \"kubernetes.io/projected/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d-kube-api-access-vqw2p\") on node \"crc\" DevicePath \"\"" Oct 02 16:58:57 crc kubenswrapper[4808]: I1002 16:58:57.926963 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.000517 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-2ws9d"] Oct 02 16:58:58 crc kubenswrapper[4808]: E1002 16:58:58.001171 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="688a2bb4-4a3a-4a13-87a8-fbf59024dd3d" containerName="neutron-httpd" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.001199 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="688a2bb4-4a3a-4a13-87a8-fbf59024dd3d" containerName="neutron-httpd" Oct 02 16:58:58 crc kubenswrapper[4808]: E1002 16:58:58.001255 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="688a2bb4-4a3a-4a13-87a8-fbf59024dd3d" containerName="neutron-api" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.001270 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="688a2bb4-4a3a-4a13-87a8-fbf59024dd3d" containerName="neutron-api" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.001493 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="688a2bb4-4a3a-4a13-87a8-fbf59024dd3d" containerName="neutron-api" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.001522 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="688a2bb4-4a3a-4a13-87a8-fbf59024dd3d" containerName="neutron-httpd" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.002528 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2ws9d" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.013879 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-2ws9d"] Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.031587 4808 generic.go:334] "Generic (PLEG): container finished" podID="688a2bb4-4a3a-4a13-87a8-fbf59024dd3d" containerID="1472f9b375176e20444d88bf3d34fe527360069ebe368cd9ad0f391466df3b5b" exitCode=0 Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.031697 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59d95756b-rl9bd" event={"ID":"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d","Type":"ContainerDied","Data":"1472f9b375176e20444d88bf3d34fe527360069ebe368cd9ad0f391466df3b5b"} Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.031735 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59d95756b-rl9bd" event={"ID":"688a2bb4-4a3a-4a13-87a8-fbf59024dd3d","Type":"ContainerDied","Data":"3874ed3d128e268bb5d1dd244afd11d6de52e6b8aeb76e8fbcb15ce335f81297"} Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.031758 4808 scope.go:117] "RemoveContainer" containerID="7da26f1db3010f3f7f174cddb5acd13403301a6470274ecef4756758ffdc9612" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.031906 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-59d95756b-rl9bd" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.042099 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56fad39c-0f30-463b-ab11-939c07877fcb","Type":"ContainerStarted","Data":"95ecfc2905202f2caccbcfd8fff58b6ecd46869268876ac62180654c94824e1b"} Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.068194 4808 scope.go:117] "RemoveContainer" containerID="1472f9b375176e20444d88bf3d34fe527360069ebe368cd9ad0f391466df3b5b" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.081452 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-59d95756b-rl9bd"] Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.092671 4808 scope.go:117] "RemoveContainer" containerID="7da26f1db3010f3f7f174cddb5acd13403301a6470274ecef4756758ffdc9612" Oct 02 16:58:58 crc kubenswrapper[4808]: E1002 16:58:58.093164 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7da26f1db3010f3f7f174cddb5acd13403301a6470274ecef4756758ffdc9612\": container with ID starting with 7da26f1db3010f3f7f174cddb5acd13403301a6470274ecef4756758ffdc9612 not found: ID does not exist" containerID="7da26f1db3010f3f7f174cddb5acd13403301a6470274ecef4756758ffdc9612" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.093203 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7da26f1db3010f3f7f174cddb5acd13403301a6470274ecef4756758ffdc9612"} err="failed to get container status \"7da26f1db3010f3f7f174cddb5acd13403301a6470274ecef4756758ffdc9612\": rpc error: code = NotFound desc = could not find container \"7da26f1db3010f3f7f174cddb5acd13403301a6470274ecef4756758ffdc9612\": container with ID starting with 7da26f1db3010f3f7f174cddb5acd13403301a6470274ecef4756758ffdc9612 not found: ID does not exist" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.093222 4808 scope.go:117] "RemoveContainer" containerID="1472f9b375176e20444d88bf3d34fe527360069ebe368cd9ad0f391466df3b5b" Oct 02 16:58:58 crc kubenswrapper[4808]: E1002 16:58:58.094006 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1472f9b375176e20444d88bf3d34fe527360069ebe368cd9ad0f391466df3b5b\": container with ID starting with 1472f9b375176e20444d88bf3d34fe527360069ebe368cd9ad0f391466df3b5b not found: ID does not exist" containerID="1472f9b375176e20444d88bf3d34fe527360069ebe368cd9ad0f391466df3b5b" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.094032 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1472f9b375176e20444d88bf3d34fe527360069ebe368cd9ad0f391466df3b5b"} err="failed to get container status \"1472f9b375176e20444d88bf3d34fe527360069ebe368cd9ad0f391466df3b5b\": rpc error: code = NotFound desc = could not find container \"1472f9b375176e20444d88bf3d34fe527360069ebe368cd9ad0f391466df3b5b\": container with ID starting with 1472f9b375176e20444d88bf3d34fe527360069ebe368cd9ad0f391466df3b5b not found: ID does not exist" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.095832 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-59d95756b-rl9bd"] Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.104999 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-x5dj9"] Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.106127 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-x5dj9" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.110951 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-x5dj9"] Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.111069 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjdb7\" (UniqueName: \"kubernetes.io/projected/91cad4f7-82d4-4b24-b77d-4e98526c6d77-kube-api-access-kjdb7\") pod \"nova-api-db-create-2ws9d\" (UID: \"91cad4f7-82d4-4b24-b77d-4e98526c6d77\") " pod="openstack/nova-api-db-create-2ws9d" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.212589 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjdb7\" (UniqueName: \"kubernetes.io/projected/91cad4f7-82d4-4b24-b77d-4e98526c6d77-kube-api-access-kjdb7\") pod \"nova-api-db-create-2ws9d\" (UID: \"91cad4f7-82d4-4b24-b77d-4e98526c6d77\") " pod="openstack/nova-api-db-create-2ws9d" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.212666 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxpzz\" (UniqueName: \"kubernetes.io/projected/1c64446b-d171-4749-a4b6-0b71c80ce9da-kube-api-access-dxpzz\") pod \"nova-cell0-db-create-x5dj9\" (UID: \"1c64446b-d171-4749-a4b6-0b71c80ce9da\") " pod="openstack/nova-cell0-db-create-x5dj9" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.228384 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjdb7\" (UniqueName: \"kubernetes.io/projected/91cad4f7-82d4-4b24-b77d-4e98526c6d77-kube-api-access-kjdb7\") pod \"nova-api-db-create-2ws9d\" (UID: \"91cad4f7-82d4-4b24-b77d-4e98526c6d77\") " pod="openstack/nova-api-db-create-2ws9d" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.299291 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-nkghg"] Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.300350 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-nkghg" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.306961 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-nkghg"] Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.314025 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxpzz\" (UniqueName: \"kubernetes.io/projected/1c64446b-d171-4749-a4b6-0b71c80ce9da-kube-api-access-dxpzz\") pod \"nova-cell0-db-create-x5dj9\" (UID: \"1c64446b-d171-4749-a4b6-0b71c80ce9da\") " pod="openstack/nova-cell0-db-create-x5dj9" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.328595 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2ws9d" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.345508 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxpzz\" (UniqueName: \"kubernetes.io/projected/1c64446b-d171-4749-a4b6-0b71c80ce9da-kube-api-access-dxpzz\") pod \"nova-cell0-db-create-x5dj9\" (UID: \"1c64446b-d171-4749-a4b6-0b71c80ce9da\") " pod="openstack/nova-cell0-db-create-x5dj9" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.415830 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mwnf\" (UniqueName: \"kubernetes.io/projected/7921b775-ed41-47c3-b334-0345e978f9d0-kube-api-access-9mwnf\") pod \"nova-cell1-db-create-nkghg\" (UID: \"7921b775-ed41-47c3-b334-0345e978f9d0\") " pod="openstack/nova-cell1-db-create-nkghg" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.430943 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-x5dj9" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.517966 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mwnf\" (UniqueName: \"kubernetes.io/projected/7921b775-ed41-47c3-b334-0345e978f9d0-kube-api-access-9mwnf\") pod \"nova-cell1-db-create-nkghg\" (UID: \"7921b775-ed41-47c3-b334-0345e978f9d0\") " pod="openstack/nova-cell1-db-create-nkghg" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.541794 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mwnf\" (UniqueName: \"kubernetes.io/projected/7921b775-ed41-47c3-b334-0345e978f9d0-kube-api-access-9mwnf\") pod \"nova-cell1-db-create-nkghg\" (UID: \"7921b775-ed41-47c3-b334-0345e978f9d0\") " pod="openstack/nova-cell1-db-create-nkghg" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.620707 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-nkghg" Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.805198 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-2ws9d"] Oct 02 16:58:58 crc kubenswrapper[4808]: W1002 16:58:58.808612 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91cad4f7_82d4_4b24_b77d_4e98526c6d77.slice/crio-f89042fccaf721ff0c5bcb59606300cc1a585d09d13c151edf03ab6a1bba31c7 WatchSource:0}: Error finding container f89042fccaf721ff0c5bcb59606300cc1a585d09d13c151edf03ab6a1bba31c7: Status 404 returned error can't find the container with id f89042fccaf721ff0c5bcb59606300cc1a585d09d13c151edf03ab6a1bba31c7 Oct 02 16:58:58 crc kubenswrapper[4808]: I1002 16:58:58.948273 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-x5dj9"] Oct 02 16:58:59 crc kubenswrapper[4808]: I1002 16:58:59.051056 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56fad39c-0f30-463b-ab11-939c07877fcb","Type":"ContainerStarted","Data":"d2e9604c0ac812c2f667a14b5c1080da4883980666213fadac46ba641cf3d480"} Oct 02 16:58:59 crc kubenswrapper[4808]: I1002 16:58:59.053222 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-2ws9d" event={"ID":"91cad4f7-82d4-4b24-b77d-4e98526c6d77","Type":"ContainerStarted","Data":"047213d1e0e0840b8ac9ea10d6144e60fbe9035487013711178ba6927959acad"} Oct 02 16:58:59 crc kubenswrapper[4808]: I1002 16:58:59.053272 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-2ws9d" event={"ID":"91cad4f7-82d4-4b24-b77d-4e98526c6d77","Type":"ContainerStarted","Data":"f89042fccaf721ff0c5bcb59606300cc1a585d09d13c151edf03ab6a1bba31c7"} Oct 02 16:58:59 crc kubenswrapper[4808]: I1002 16:58:59.055361 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-x5dj9" event={"ID":"1c64446b-d171-4749-a4b6-0b71c80ce9da","Type":"ContainerStarted","Data":"99b9e1c562350f53642c0f73425df96b878524bdf17b5e1a5692ba536ac429a6"} Oct 02 16:58:59 crc kubenswrapper[4808]: I1002 16:58:59.070125 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-2ws9d" podStartSLOduration=2.070107786 podStartE2EDuration="2.070107786s" podCreationTimestamp="2025-10-02 16:58:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:58:59.065584493 +0000 UTC m=+1106.391113493" watchObservedRunningTime="2025-10-02 16:58:59.070107786 +0000 UTC m=+1106.395636786" Oct 02 16:58:59 crc kubenswrapper[4808]: I1002 16:58:59.114860 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-nkghg"] Oct 02 16:58:59 crc kubenswrapper[4808]: W1002 16:58:59.138452 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7921b775_ed41_47c3_b334_0345e978f9d0.slice/crio-a428d7f0c9e813c1999381c4a7275cac9dd78a82fb15b4c900c181ac98ed31f7 WatchSource:0}: Error finding container a428d7f0c9e813c1999381c4a7275cac9dd78a82fb15b4c900c181ac98ed31f7: Status 404 returned error can't find the container with id a428d7f0c9e813c1999381c4a7275cac9dd78a82fb15b4c900c181ac98ed31f7 Oct 02 16:58:59 crc kubenswrapper[4808]: I1002 16:58:59.412482 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="688a2bb4-4a3a-4a13-87a8-fbf59024dd3d" path="/var/lib/kubelet/pods/688a2bb4-4a3a-4a13-87a8-fbf59024dd3d/volumes" Oct 02 16:59:00 crc kubenswrapper[4808]: I1002 16:59:00.064227 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56fad39c-0f30-463b-ab11-939c07877fcb","Type":"ContainerStarted","Data":"a1df72d17ada398d2731f64c8e696ef9b0e8f041c61bc71a5ae71fe44830a097"} Oct 02 16:59:00 crc kubenswrapper[4808]: I1002 16:59:00.065582 4808 generic.go:334] "Generic (PLEG): container finished" podID="91cad4f7-82d4-4b24-b77d-4e98526c6d77" containerID="047213d1e0e0840b8ac9ea10d6144e60fbe9035487013711178ba6927959acad" exitCode=0 Oct 02 16:59:00 crc kubenswrapper[4808]: I1002 16:59:00.065631 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-2ws9d" event={"ID":"91cad4f7-82d4-4b24-b77d-4e98526c6d77","Type":"ContainerDied","Data":"047213d1e0e0840b8ac9ea10d6144e60fbe9035487013711178ba6927959acad"} Oct 02 16:59:00 crc kubenswrapper[4808]: I1002 16:59:00.066799 4808 generic.go:334] "Generic (PLEG): container finished" podID="1c64446b-d171-4749-a4b6-0b71c80ce9da" containerID="bd67e11c8060c1fe134a3721e2c51c04da63cbb6ea023b09aeaf6b6ea1b10f3a" exitCode=0 Oct 02 16:59:00 crc kubenswrapper[4808]: I1002 16:59:00.066845 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-x5dj9" event={"ID":"1c64446b-d171-4749-a4b6-0b71c80ce9da","Type":"ContainerDied","Data":"bd67e11c8060c1fe134a3721e2c51c04da63cbb6ea023b09aeaf6b6ea1b10f3a"} Oct 02 16:59:00 crc kubenswrapper[4808]: I1002 16:59:00.068844 4808 generic.go:334] "Generic (PLEG): container finished" podID="7921b775-ed41-47c3-b334-0345e978f9d0" containerID="e1c9417e4f2dc43ca95e4a3755f8d9316ee5b61c0677dda5c6e20ebeb9efdbc3" exitCode=0 Oct 02 16:59:00 crc kubenswrapper[4808]: I1002 16:59:00.068888 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-nkghg" event={"ID":"7921b775-ed41-47c3-b334-0345e978f9d0","Type":"ContainerDied","Data":"e1c9417e4f2dc43ca95e4a3755f8d9316ee5b61c0677dda5c6e20ebeb9efdbc3"} Oct 02 16:59:00 crc kubenswrapper[4808]: I1002 16:59:00.068915 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-nkghg" event={"ID":"7921b775-ed41-47c3-b334-0345e978f9d0","Type":"ContainerStarted","Data":"a428d7f0c9e813c1999381c4a7275cac9dd78a82fb15b4c900c181ac98ed31f7"} Oct 02 16:59:00 crc kubenswrapper[4808]: I1002 16:59:00.345348 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:59:01 crc kubenswrapper[4808]: I1002 16:59:01.081412 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56fad39c-0f30-463b-ab11-939c07877fcb","Type":"ContainerStarted","Data":"607c8efb177060a6c37a979612e33b967654da2d8ae229ddac8db0c88b106d0c"} Oct 02 16:59:01 crc kubenswrapper[4808]: I1002 16:59:01.598174 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2ws9d" Oct 02 16:59:01 crc kubenswrapper[4808]: I1002 16:59:01.604616 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-x5dj9" Oct 02 16:59:01 crc kubenswrapper[4808]: I1002 16:59:01.617448 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-nkghg" Oct 02 16:59:01 crc kubenswrapper[4808]: I1002 16:59:01.680592 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxpzz\" (UniqueName: \"kubernetes.io/projected/1c64446b-d171-4749-a4b6-0b71c80ce9da-kube-api-access-dxpzz\") pod \"1c64446b-d171-4749-a4b6-0b71c80ce9da\" (UID: \"1c64446b-d171-4749-a4b6-0b71c80ce9da\") " Oct 02 16:59:01 crc kubenswrapper[4808]: I1002 16:59:01.680751 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mwnf\" (UniqueName: \"kubernetes.io/projected/7921b775-ed41-47c3-b334-0345e978f9d0-kube-api-access-9mwnf\") pod \"7921b775-ed41-47c3-b334-0345e978f9d0\" (UID: \"7921b775-ed41-47c3-b334-0345e978f9d0\") " Oct 02 16:59:01 crc kubenswrapper[4808]: I1002 16:59:01.680853 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjdb7\" (UniqueName: \"kubernetes.io/projected/91cad4f7-82d4-4b24-b77d-4e98526c6d77-kube-api-access-kjdb7\") pod \"91cad4f7-82d4-4b24-b77d-4e98526c6d77\" (UID: \"91cad4f7-82d4-4b24-b77d-4e98526c6d77\") " Oct 02 16:59:01 crc kubenswrapper[4808]: I1002 16:59:01.686318 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c64446b-d171-4749-a4b6-0b71c80ce9da-kube-api-access-dxpzz" (OuterVolumeSpecName: "kube-api-access-dxpzz") pod "1c64446b-d171-4749-a4b6-0b71c80ce9da" (UID: "1c64446b-d171-4749-a4b6-0b71c80ce9da"). InnerVolumeSpecName "kube-api-access-dxpzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:59:01 crc kubenswrapper[4808]: I1002 16:59:01.695772 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91cad4f7-82d4-4b24-b77d-4e98526c6d77-kube-api-access-kjdb7" (OuterVolumeSpecName: "kube-api-access-kjdb7") pod "91cad4f7-82d4-4b24-b77d-4e98526c6d77" (UID: "91cad4f7-82d4-4b24-b77d-4e98526c6d77"). InnerVolumeSpecName "kube-api-access-kjdb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:59:01 crc kubenswrapper[4808]: I1002 16:59:01.695838 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7921b775-ed41-47c3-b334-0345e978f9d0-kube-api-access-9mwnf" (OuterVolumeSpecName: "kube-api-access-9mwnf") pod "7921b775-ed41-47c3-b334-0345e978f9d0" (UID: "7921b775-ed41-47c3-b334-0345e978f9d0"). InnerVolumeSpecName "kube-api-access-9mwnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:59:01 crc kubenswrapper[4808]: I1002 16:59:01.783098 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxpzz\" (UniqueName: \"kubernetes.io/projected/1c64446b-d171-4749-a4b6-0b71c80ce9da-kube-api-access-dxpzz\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:01 crc kubenswrapper[4808]: I1002 16:59:01.783138 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mwnf\" (UniqueName: \"kubernetes.io/projected/7921b775-ed41-47c3-b334-0345e978f9d0-kube-api-access-9mwnf\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:01 crc kubenswrapper[4808]: I1002 16:59:01.783148 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjdb7\" (UniqueName: \"kubernetes.io/projected/91cad4f7-82d4-4b24-b77d-4e98526c6d77-kube-api-access-kjdb7\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:02 crc kubenswrapper[4808]: I1002 16:59:02.094911 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-x5dj9" Oct 02 16:59:02 crc kubenswrapper[4808]: I1002 16:59:02.094911 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-x5dj9" event={"ID":"1c64446b-d171-4749-a4b6-0b71c80ce9da","Type":"ContainerDied","Data":"99b9e1c562350f53642c0f73425df96b878524bdf17b5e1a5692ba536ac429a6"} Oct 02 16:59:02 crc kubenswrapper[4808]: I1002 16:59:02.095477 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99b9e1c562350f53642c0f73425df96b878524bdf17b5e1a5692ba536ac429a6" Oct 02 16:59:02 crc kubenswrapper[4808]: I1002 16:59:02.096784 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-nkghg" event={"ID":"7921b775-ed41-47c3-b334-0345e978f9d0","Type":"ContainerDied","Data":"a428d7f0c9e813c1999381c4a7275cac9dd78a82fb15b4c900c181ac98ed31f7"} Oct 02 16:59:02 crc kubenswrapper[4808]: I1002 16:59:02.096821 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a428d7f0c9e813c1999381c4a7275cac9dd78a82fb15b4c900c181ac98ed31f7" Oct 02 16:59:02 crc kubenswrapper[4808]: I1002 16:59:02.097318 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-nkghg" Oct 02 16:59:02 crc kubenswrapper[4808]: I1002 16:59:02.098536 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-2ws9d" event={"ID":"91cad4f7-82d4-4b24-b77d-4e98526c6d77","Type":"ContainerDied","Data":"f89042fccaf721ff0c5bcb59606300cc1a585d09d13c151edf03ab6a1bba31c7"} Oct 02 16:59:02 crc kubenswrapper[4808]: I1002 16:59:02.098579 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f89042fccaf721ff0c5bcb59606300cc1a585d09d13c151edf03ab6a1bba31c7" Oct 02 16:59:02 crc kubenswrapper[4808]: I1002 16:59:02.098588 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2ws9d" Oct 02 16:59:03 crc kubenswrapper[4808]: I1002 16:59:03.109469 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56fad39c-0f30-463b-ab11-939c07877fcb","Type":"ContainerStarted","Data":"afc5bee5f25acaf633aebfd16fb789f0ef10ffb5a59a7b16bc3f46ea489cfec6"} Oct 02 16:59:03 crc kubenswrapper[4808]: I1002 16:59:03.109840 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 16:59:03 crc kubenswrapper[4808]: I1002 16:59:03.109742 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="56fad39c-0f30-463b-ab11-939c07877fcb" containerName="ceilometer-notification-agent" containerID="cri-o://a1df72d17ada398d2731f64c8e696ef9b0e8f041c61bc71a5ae71fe44830a097" gracePeriod=30 Oct 02 16:59:03 crc kubenswrapper[4808]: I1002 16:59:03.109702 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="56fad39c-0f30-463b-ab11-939c07877fcb" containerName="ceilometer-central-agent" containerID="cri-o://d2e9604c0ac812c2f667a14b5c1080da4883980666213fadac46ba641cf3d480" gracePeriod=30 Oct 02 16:59:03 crc kubenswrapper[4808]: I1002 16:59:03.109734 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="56fad39c-0f30-463b-ab11-939c07877fcb" containerName="sg-core" containerID="cri-o://607c8efb177060a6c37a979612e33b967654da2d8ae229ddac8db0c88b106d0c" gracePeriod=30 Oct 02 16:59:03 crc kubenswrapper[4808]: I1002 16:59:03.109772 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="56fad39c-0f30-463b-ab11-939c07877fcb" containerName="proxy-httpd" containerID="cri-o://afc5bee5f25acaf633aebfd16fb789f0ef10ffb5a59a7b16bc3f46ea489cfec6" gracePeriod=30 Oct 02 16:59:03 crc kubenswrapper[4808]: I1002 16:59:03.145939 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.868140913 podStartE2EDuration="6.145916961s" podCreationTimestamp="2025-10-02 16:58:57 +0000 UTC" firstStartedPulling="2025-10-02 16:58:57.931614109 +0000 UTC m=+1105.257143109" lastFinishedPulling="2025-10-02 16:59:02.209390147 +0000 UTC m=+1109.534919157" observedRunningTime="2025-10-02 16:59:03.14405128 +0000 UTC m=+1110.469580280" watchObservedRunningTime="2025-10-02 16:59:03.145916961 +0000 UTC m=+1110.471445961" Oct 02 16:59:04 crc kubenswrapper[4808]: I1002 16:59:04.119620 4808 generic.go:334] "Generic (PLEG): container finished" podID="56fad39c-0f30-463b-ab11-939c07877fcb" containerID="afc5bee5f25acaf633aebfd16fb789f0ef10ffb5a59a7b16bc3f46ea489cfec6" exitCode=0 Oct 02 16:59:04 crc kubenswrapper[4808]: I1002 16:59:04.119657 4808 generic.go:334] "Generic (PLEG): container finished" podID="56fad39c-0f30-463b-ab11-939c07877fcb" containerID="607c8efb177060a6c37a979612e33b967654da2d8ae229ddac8db0c88b106d0c" exitCode=2 Oct 02 16:59:04 crc kubenswrapper[4808]: I1002 16:59:04.119669 4808 generic.go:334] "Generic (PLEG): container finished" podID="56fad39c-0f30-463b-ab11-939c07877fcb" containerID="a1df72d17ada398d2731f64c8e696ef9b0e8f041c61bc71a5ae71fe44830a097" exitCode=0 Oct 02 16:59:04 crc kubenswrapper[4808]: I1002 16:59:04.119702 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56fad39c-0f30-463b-ab11-939c07877fcb","Type":"ContainerDied","Data":"afc5bee5f25acaf633aebfd16fb789f0ef10ffb5a59a7b16bc3f46ea489cfec6"} Oct 02 16:59:04 crc kubenswrapper[4808]: I1002 16:59:04.119747 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56fad39c-0f30-463b-ab11-939c07877fcb","Type":"ContainerDied","Data":"607c8efb177060a6c37a979612e33b967654da2d8ae229ddac8db0c88b106d0c"} Oct 02 16:59:04 crc kubenswrapper[4808]: I1002 16:59:04.119762 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56fad39c-0f30-463b-ab11-939c07877fcb","Type":"ContainerDied","Data":"a1df72d17ada398d2731f64c8e696ef9b0e8f041c61bc71a5ae71fe44830a097"} Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.601727 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.645951 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-scripts\") pod \"56fad39c-0f30-463b-ab11-939c07877fcb\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.646125 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g44tx\" (UniqueName: \"kubernetes.io/projected/56fad39c-0f30-463b-ab11-939c07877fcb-kube-api-access-g44tx\") pod \"56fad39c-0f30-463b-ab11-939c07877fcb\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.646156 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-config-data\") pod \"56fad39c-0f30-463b-ab11-939c07877fcb\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.646179 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-sg-core-conf-yaml\") pod \"56fad39c-0f30-463b-ab11-939c07877fcb\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.646225 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56fad39c-0f30-463b-ab11-939c07877fcb-run-httpd\") pod \"56fad39c-0f30-463b-ab11-939c07877fcb\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.646288 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56fad39c-0f30-463b-ab11-939c07877fcb-log-httpd\") pod \"56fad39c-0f30-463b-ab11-939c07877fcb\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.646358 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-combined-ca-bundle\") pod \"56fad39c-0f30-463b-ab11-939c07877fcb\" (UID: \"56fad39c-0f30-463b-ab11-939c07877fcb\") " Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.652674 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56fad39c-0f30-463b-ab11-939c07877fcb-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "56fad39c-0f30-463b-ab11-939c07877fcb" (UID: "56fad39c-0f30-463b-ab11-939c07877fcb"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.652791 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56fad39c-0f30-463b-ab11-939c07877fcb-kube-api-access-g44tx" (OuterVolumeSpecName: "kube-api-access-g44tx") pod "56fad39c-0f30-463b-ab11-939c07877fcb" (UID: "56fad39c-0f30-463b-ab11-939c07877fcb"). InnerVolumeSpecName "kube-api-access-g44tx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.653058 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56fad39c-0f30-463b-ab11-939c07877fcb-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "56fad39c-0f30-463b-ab11-939c07877fcb" (UID: "56fad39c-0f30-463b-ab11-939c07877fcb"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.660306 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-scripts" (OuterVolumeSpecName: "scripts") pod "56fad39c-0f30-463b-ab11-939c07877fcb" (UID: "56fad39c-0f30-463b-ab11-939c07877fcb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.679992 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "56fad39c-0f30-463b-ab11-939c07877fcb" (UID: "56fad39c-0f30-463b-ab11-939c07877fcb"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.743923 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56fad39c-0f30-463b-ab11-939c07877fcb" (UID: "56fad39c-0f30-463b-ab11-939c07877fcb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.748909 4808 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56fad39c-0f30-463b-ab11-939c07877fcb-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.748941 4808 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56fad39c-0f30-463b-ab11-939c07877fcb-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.748950 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.748962 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.748971 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g44tx\" (UniqueName: \"kubernetes.io/projected/56fad39c-0f30-463b-ab11-939c07877fcb-kube-api-access-g44tx\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.748982 4808 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.750629 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-config-data" (OuterVolumeSpecName: "config-data") pod "56fad39c-0f30-463b-ab11-939c07877fcb" (UID: "56fad39c-0f30-463b-ab11-939c07877fcb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:05 crc kubenswrapper[4808]: I1002 16:59:05.849932 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56fad39c-0f30-463b-ab11-939c07877fcb-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.137320 4808 generic.go:334] "Generic (PLEG): container finished" podID="56fad39c-0f30-463b-ab11-939c07877fcb" containerID="d2e9604c0ac812c2f667a14b5c1080da4883980666213fadac46ba641cf3d480" exitCode=0 Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.137369 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56fad39c-0f30-463b-ab11-939c07877fcb","Type":"ContainerDied","Data":"d2e9604c0ac812c2f667a14b5c1080da4883980666213fadac46ba641cf3d480"} Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.137399 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56fad39c-0f30-463b-ab11-939c07877fcb","Type":"ContainerDied","Data":"95ecfc2905202f2caccbcfd8fff58b6ecd46869268876ac62180654c94824e1b"} Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.137437 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.137442 4808 scope.go:117] "RemoveContainer" containerID="afc5bee5f25acaf633aebfd16fb789f0ef10ffb5a59a7b16bc3f46ea489cfec6" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.192428 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.199708 4808 scope.go:117] "RemoveContainer" containerID="607c8efb177060a6c37a979612e33b967654da2d8ae229ddac8db0c88b106d0c" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.206761 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.227221 4808 scope.go:117] "RemoveContainer" containerID="a1df72d17ada398d2731f64c8e696ef9b0e8f041c61bc71a5ae71fe44830a097" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.262168 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:59:06 crc kubenswrapper[4808]: E1002 16:59:06.263051 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56fad39c-0f30-463b-ab11-939c07877fcb" containerName="sg-core" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.263081 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="56fad39c-0f30-463b-ab11-939c07877fcb" containerName="sg-core" Oct 02 16:59:06 crc kubenswrapper[4808]: E1002 16:59:06.263106 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91cad4f7-82d4-4b24-b77d-4e98526c6d77" containerName="mariadb-database-create" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.263119 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="91cad4f7-82d4-4b24-b77d-4e98526c6d77" containerName="mariadb-database-create" Oct 02 16:59:06 crc kubenswrapper[4808]: E1002 16:59:06.263151 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c64446b-d171-4749-a4b6-0b71c80ce9da" containerName="mariadb-database-create" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.263165 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c64446b-d171-4749-a4b6-0b71c80ce9da" containerName="mariadb-database-create" Oct 02 16:59:06 crc kubenswrapper[4808]: E1002 16:59:06.263230 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56fad39c-0f30-463b-ab11-939c07877fcb" containerName="ceilometer-notification-agent" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.263266 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="56fad39c-0f30-463b-ab11-939c07877fcb" containerName="ceilometer-notification-agent" Oct 02 16:59:06 crc kubenswrapper[4808]: E1002 16:59:06.263287 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7921b775-ed41-47c3-b334-0345e978f9d0" containerName="mariadb-database-create" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.263299 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7921b775-ed41-47c3-b334-0345e978f9d0" containerName="mariadb-database-create" Oct 02 16:59:06 crc kubenswrapper[4808]: E1002 16:59:06.263313 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56fad39c-0f30-463b-ab11-939c07877fcb" containerName="proxy-httpd" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.263325 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="56fad39c-0f30-463b-ab11-939c07877fcb" containerName="proxy-httpd" Oct 02 16:59:06 crc kubenswrapper[4808]: E1002 16:59:06.263349 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56fad39c-0f30-463b-ab11-939c07877fcb" containerName="ceilometer-central-agent" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.263361 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="56fad39c-0f30-463b-ab11-939c07877fcb" containerName="ceilometer-central-agent" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.263688 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="91cad4f7-82d4-4b24-b77d-4e98526c6d77" containerName="mariadb-database-create" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.263719 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c64446b-d171-4749-a4b6-0b71c80ce9da" containerName="mariadb-database-create" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.263742 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7921b775-ed41-47c3-b334-0345e978f9d0" containerName="mariadb-database-create" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.263772 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="56fad39c-0f30-463b-ab11-939c07877fcb" containerName="ceilometer-notification-agent" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.263792 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="56fad39c-0f30-463b-ab11-939c07877fcb" containerName="proxy-httpd" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.263817 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="56fad39c-0f30-463b-ab11-939c07877fcb" containerName="sg-core" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.263842 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="56fad39c-0f30-463b-ab11-939c07877fcb" containerName="ceilometer-central-agent" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.265101 4808 scope.go:117] "RemoveContainer" containerID="d2e9604c0ac812c2f667a14b5c1080da4883980666213fadac46ba641cf3d480" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.266982 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.272009 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.272199 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.277892 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.287604 4808 scope.go:117] "RemoveContainer" containerID="afc5bee5f25acaf633aebfd16fb789f0ef10ffb5a59a7b16bc3f46ea489cfec6" Oct 02 16:59:06 crc kubenswrapper[4808]: E1002 16:59:06.289660 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afc5bee5f25acaf633aebfd16fb789f0ef10ffb5a59a7b16bc3f46ea489cfec6\": container with ID starting with afc5bee5f25acaf633aebfd16fb789f0ef10ffb5a59a7b16bc3f46ea489cfec6 not found: ID does not exist" containerID="afc5bee5f25acaf633aebfd16fb789f0ef10ffb5a59a7b16bc3f46ea489cfec6" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.289699 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afc5bee5f25acaf633aebfd16fb789f0ef10ffb5a59a7b16bc3f46ea489cfec6"} err="failed to get container status \"afc5bee5f25acaf633aebfd16fb789f0ef10ffb5a59a7b16bc3f46ea489cfec6\": rpc error: code = NotFound desc = could not find container \"afc5bee5f25acaf633aebfd16fb789f0ef10ffb5a59a7b16bc3f46ea489cfec6\": container with ID starting with afc5bee5f25acaf633aebfd16fb789f0ef10ffb5a59a7b16bc3f46ea489cfec6 not found: ID does not exist" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.289725 4808 scope.go:117] "RemoveContainer" containerID="607c8efb177060a6c37a979612e33b967654da2d8ae229ddac8db0c88b106d0c" Oct 02 16:59:06 crc kubenswrapper[4808]: E1002 16:59:06.289973 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"607c8efb177060a6c37a979612e33b967654da2d8ae229ddac8db0c88b106d0c\": container with ID starting with 607c8efb177060a6c37a979612e33b967654da2d8ae229ddac8db0c88b106d0c not found: ID does not exist" containerID="607c8efb177060a6c37a979612e33b967654da2d8ae229ddac8db0c88b106d0c" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.289997 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"607c8efb177060a6c37a979612e33b967654da2d8ae229ddac8db0c88b106d0c"} err="failed to get container status \"607c8efb177060a6c37a979612e33b967654da2d8ae229ddac8db0c88b106d0c\": rpc error: code = NotFound desc = could not find container \"607c8efb177060a6c37a979612e33b967654da2d8ae229ddac8db0c88b106d0c\": container with ID starting with 607c8efb177060a6c37a979612e33b967654da2d8ae229ddac8db0c88b106d0c not found: ID does not exist" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.290015 4808 scope.go:117] "RemoveContainer" containerID="a1df72d17ada398d2731f64c8e696ef9b0e8f041c61bc71a5ae71fe44830a097" Oct 02 16:59:06 crc kubenswrapper[4808]: E1002 16:59:06.290268 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1df72d17ada398d2731f64c8e696ef9b0e8f041c61bc71a5ae71fe44830a097\": container with ID starting with a1df72d17ada398d2731f64c8e696ef9b0e8f041c61bc71a5ae71fe44830a097 not found: ID does not exist" containerID="a1df72d17ada398d2731f64c8e696ef9b0e8f041c61bc71a5ae71fe44830a097" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.290291 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1df72d17ada398d2731f64c8e696ef9b0e8f041c61bc71a5ae71fe44830a097"} err="failed to get container status \"a1df72d17ada398d2731f64c8e696ef9b0e8f041c61bc71a5ae71fe44830a097\": rpc error: code = NotFound desc = could not find container \"a1df72d17ada398d2731f64c8e696ef9b0e8f041c61bc71a5ae71fe44830a097\": container with ID starting with a1df72d17ada398d2731f64c8e696ef9b0e8f041c61bc71a5ae71fe44830a097 not found: ID does not exist" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.290308 4808 scope.go:117] "RemoveContainer" containerID="d2e9604c0ac812c2f667a14b5c1080da4883980666213fadac46ba641cf3d480" Oct 02 16:59:06 crc kubenswrapper[4808]: E1002 16:59:06.290533 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2e9604c0ac812c2f667a14b5c1080da4883980666213fadac46ba641cf3d480\": container with ID starting with d2e9604c0ac812c2f667a14b5c1080da4883980666213fadac46ba641cf3d480 not found: ID does not exist" containerID="d2e9604c0ac812c2f667a14b5c1080da4883980666213fadac46ba641cf3d480" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.290554 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2e9604c0ac812c2f667a14b5c1080da4883980666213fadac46ba641cf3d480"} err="failed to get container status \"d2e9604c0ac812c2f667a14b5c1080da4883980666213fadac46ba641cf3d480\": rpc error: code = NotFound desc = could not find container \"d2e9604c0ac812c2f667a14b5c1080da4883980666213fadac46ba641cf3d480\": container with ID starting with d2e9604c0ac812c2f667a14b5c1080da4883980666213fadac46ba641cf3d480 not found: ID does not exist" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.358748 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.358827 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsckg\" (UniqueName: \"kubernetes.io/projected/e31a6648-4159-4434-ac4f-7f5c142a650b-kube-api-access-nsckg\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.358846 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-scripts\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.358861 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.358931 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e31a6648-4159-4434-ac4f-7f5c142a650b-log-httpd\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.358953 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-config-data\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.358975 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e31a6648-4159-4434-ac4f-7f5c142a650b-run-httpd\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.461010 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e31a6648-4159-4434-ac4f-7f5c142a650b-log-httpd\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.461095 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-config-data\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.461157 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e31a6648-4159-4434-ac4f-7f5c142a650b-run-httpd\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.461227 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.461396 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsckg\" (UniqueName: \"kubernetes.io/projected/e31a6648-4159-4434-ac4f-7f5c142a650b-kube-api-access-nsckg\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.461508 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-scripts\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.461553 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.463367 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e31a6648-4159-4434-ac4f-7f5c142a650b-log-httpd\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.463430 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e31a6648-4159-4434-ac4f-7f5c142a650b-run-httpd\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.466857 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-scripts\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.467284 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.468203 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-config-data\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.468408 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.492769 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsckg\" (UniqueName: \"kubernetes.io/projected/e31a6648-4159-4434-ac4f-7f5c142a650b-kube-api-access-nsckg\") pod \"ceilometer-0\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " pod="openstack/ceilometer-0" Oct 02 16:59:06 crc kubenswrapper[4808]: I1002 16:59:06.586057 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:59:07 crc kubenswrapper[4808]: I1002 16:59:07.059554 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:59:07 crc kubenswrapper[4808]: W1002 16:59:07.072427 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode31a6648_4159_4434_ac4f_7f5c142a650b.slice/crio-43e09574ae7a5a69885d462981ddb8b646c33bd95fbf171665817b3894a865e1 WatchSource:0}: Error finding container 43e09574ae7a5a69885d462981ddb8b646c33bd95fbf171665817b3894a865e1: Status 404 returned error can't find the container with id 43e09574ae7a5a69885d462981ddb8b646c33bd95fbf171665817b3894a865e1 Oct 02 16:59:07 crc kubenswrapper[4808]: I1002 16:59:07.149995 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e31a6648-4159-4434-ac4f-7f5c142a650b","Type":"ContainerStarted","Data":"43e09574ae7a5a69885d462981ddb8b646c33bd95fbf171665817b3894a865e1"} Oct 02 16:59:07 crc kubenswrapper[4808]: I1002 16:59:07.421947 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56fad39c-0f30-463b-ab11-939c07877fcb" path="/var/lib/kubelet/pods/56fad39c-0f30-463b-ab11-939c07877fcb/volumes" Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.161803 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e31a6648-4159-4434-ac4f-7f5c142a650b","Type":"ContainerStarted","Data":"2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e"} Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.228633 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-5e0c-account-create-75q46"] Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.229916 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5e0c-account-create-75q46" Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.232007 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.245015 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5e0c-account-create-75q46"] Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.293059 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5fdj\" (UniqueName: \"kubernetes.io/projected/5a2cb272-0bff-49dc-bf02-114dcfda4aea-kube-api-access-n5fdj\") pod \"nova-api-5e0c-account-create-75q46\" (UID: \"5a2cb272-0bff-49dc-bf02-114dcfda4aea\") " pod="openstack/nova-api-5e0c-account-create-75q46" Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.394307 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5fdj\" (UniqueName: \"kubernetes.io/projected/5a2cb272-0bff-49dc-bf02-114dcfda4aea-kube-api-access-n5fdj\") pod \"nova-api-5e0c-account-create-75q46\" (UID: \"5a2cb272-0bff-49dc-bf02-114dcfda4aea\") " pod="openstack/nova-api-5e0c-account-create-75q46" Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.416176 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5fdj\" (UniqueName: \"kubernetes.io/projected/5a2cb272-0bff-49dc-bf02-114dcfda4aea-kube-api-access-n5fdj\") pod \"nova-api-5e0c-account-create-75q46\" (UID: \"5a2cb272-0bff-49dc-bf02-114dcfda4aea\") " pod="openstack/nova-api-5e0c-account-create-75q46" Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.430747 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-731b-account-create-zrkfk"] Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.432052 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-731b-account-create-zrkfk" Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.438699 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.482059 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-731b-account-create-zrkfk"] Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.496145 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6szx\" (UniqueName: \"kubernetes.io/projected/e94ceeb9-6ff7-4fcd-8727-64bb034bef2f-kube-api-access-s6szx\") pod \"nova-cell0-731b-account-create-zrkfk\" (UID: \"e94ceeb9-6ff7-4fcd-8727-64bb034bef2f\") " pod="openstack/nova-cell0-731b-account-create-zrkfk" Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.545614 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5e0c-account-create-75q46" Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.610149 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6szx\" (UniqueName: \"kubernetes.io/projected/e94ceeb9-6ff7-4fcd-8727-64bb034bef2f-kube-api-access-s6szx\") pod \"nova-cell0-731b-account-create-zrkfk\" (UID: \"e94ceeb9-6ff7-4fcd-8727-64bb034bef2f\") " pod="openstack/nova-cell0-731b-account-create-zrkfk" Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.641608 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6szx\" (UniqueName: \"kubernetes.io/projected/e94ceeb9-6ff7-4fcd-8727-64bb034bef2f-kube-api-access-s6szx\") pod \"nova-cell0-731b-account-create-zrkfk\" (UID: \"e94ceeb9-6ff7-4fcd-8727-64bb034bef2f\") " pod="openstack/nova-cell0-731b-account-create-zrkfk" Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.666283 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-4976-account-create-lskbv"] Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.667373 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4976-account-create-lskbv" Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.669931 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-4976-account-create-lskbv"] Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.670059 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.711369 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqwg6\" (UniqueName: \"kubernetes.io/projected/b8311380-1e59-4d6c-b48f-c7dac382f325-kube-api-access-fqwg6\") pod \"nova-cell1-4976-account-create-lskbv\" (UID: \"b8311380-1e59-4d6c-b48f-c7dac382f325\") " pod="openstack/nova-cell1-4976-account-create-lskbv" Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.783334 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-731b-account-create-zrkfk" Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.813676 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqwg6\" (UniqueName: \"kubernetes.io/projected/b8311380-1e59-4d6c-b48f-c7dac382f325-kube-api-access-fqwg6\") pod \"nova-cell1-4976-account-create-lskbv\" (UID: \"b8311380-1e59-4d6c-b48f-c7dac382f325\") " pod="openstack/nova-cell1-4976-account-create-lskbv" Oct 02 16:59:08 crc kubenswrapper[4808]: I1002 16:59:08.830951 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqwg6\" (UniqueName: \"kubernetes.io/projected/b8311380-1e59-4d6c-b48f-c7dac382f325-kube-api-access-fqwg6\") pod \"nova-cell1-4976-account-create-lskbv\" (UID: \"b8311380-1e59-4d6c-b48f-c7dac382f325\") " pod="openstack/nova-cell1-4976-account-create-lskbv" Oct 02 16:59:09 crc kubenswrapper[4808]: I1002 16:59:09.012688 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-731b-account-create-zrkfk"] Oct 02 16:59:09 crc kubenswrapper[4808]: W1002 16:59:09.017426 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode94ceeb9_6ff7_4fcd_8727_64bb034bef2f.slice/crio-6633c2bcf65b623a175b29d52e1463fe4acd1311f7934d48ba3bb6b67ae4002d WatchSource:0}: Error finding container 6633c2bcf65b623a175b29d52e1463fe4acd1311f7934d48ba3bb6b67ae4002d: Status 404 returned error can't find the container with id 6633c2bcf65b623a175b29d52e1463fe4acd1311f7934d48ba3bb6b67ae4002d Oct 02 16:59:09 crc kubenswrapper[4808]: I1002 16:59:09.017979 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4976-account-create-lskbv" Oct 02 16:59:09 crc kubenswrapper[4808]: I1002 16:59:09.130899 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5e0c-account-create-75q46"] Oct 02 16:59:09 crc kubenswrapper[4808]: W1002 16:59:09.143448 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a2cb272_0bff_49dc_bf02_114dcfda4aea.slice/crio-a95fa1e3fe206618ea024cf0066feb2d34bdae8aea2b03cde79a6639e048d273 WatchSource:0}: Error finding container a95fa1e3fe206618ea024cf0066feb2d34bdae8aea2b03cde79a6639e048d273: Status 404 returned error can't find the container with id a95fa1e3fe206618ea024cf0066feb2d34bdae8aea2b03cde79a6639e048d273 Oct 02 16:59:09 crc kubenswrapper[4808]: I1002 16:59:09.175443 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-731b-account-create-zrkfk" event={"ID":"e94ceeb9-6ff7-4fcd-8727-64bb034bef2f","Type":"ContainerStarted","Data":"6633c2bcf65b623a175b29d52e1463fe4acd1311f7934d48ba3bb6b67ae4002d"} Oct 02 16:59:09 crc kubenswrapper[4808]: I1002 16:59:09.177394 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5e0c-account-create-75q46" event={"ID":"5a2cb272-0bff-49dc-bf02-114dcfda4aea","Type":"ContainerStarted","Data":"a95fa1e3fe206618ea024cf0066feb2d34bdae8aea2b03cde79a6639e048d273"} Oct 02 16:59:09 crc kubenswrapper[4808]: I1002 16:59:09.180456 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e31a6648-4159-4434-ac4f-7f5c142a650b","Type":"ContainerStarted","Data":"4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e"} Oct 02 16:59:09 crc kubenswrapper[4808]: I1002 16:59:09.488607 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-4976-account-create-lskbv"] Oct 02 16:59:10 crc kubenswrapper[4808]: I1002 16:59:10.194323 4808 generic.go:334] "Generic (PLEG): container finished" podID="b8311380-1e59-4d6c-b48f-c7dac382f325" containerID="9febe6f90b26edcc63c1ec945e665d6ef621b9b737b6428b9e16ac75e6bc8dba" exitCode=0 Oct 02 16:59:10 crc kubenswrapper[4808]: I1002 16:59:10.194419 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4976-account-create-lskbv" event={"ID":"b8311380-1e59-4d6c-b48f-c7dac382f325","Type":"ContainerDied","Data":"9febe6f90b26edcc63c1ec945e665d6ef621b9b737b6428b9e16ac75e6bc8dba"} Oct 02 16:59:10 crc kubenswrapper[4808]: I1002 16:59:10.194489 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4976-account-create-lskbv" event={"ID":"b8311380-1e59-4d6c-b48f-c7dac382f325","Type":"ContainerStarted","Data":"99e3a017a6dd62320985729ff487d7788f9260550d2365a17b32d92c617ce048"} Oct 02 16:59:10 crc kubenswrapper[4808]: I1002 16:59:10.196801 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e31a6648-4159-4434-ac4f-7f5c142a650b","Type":"ContainerStarted","Data":"d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9"} Oct 02 16:59:10 crc kubenswrapper[4808]: I1002 16:59:10.198965 4808 generic.go:334] "Generic (PLEG): container finished" podID="e94ceeb9-6ff7-4fcd-8727-64bb034bef2f" containerID="32750afea91cfd59fed83d3df033a4c648e3f9ac68afbd02f02fef845fc8251e" exitCode=0 Oct 02 16:59:10 crc kubenswrapper[4808]: I1002 16:59:10.199020 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-731b-account-create-zrkfk" event={"ID":"e94ceeb9-6ff7-4fcd-8727-64bb034bef2f","Type":"ContainerDied","Data":"32750afea91cfd59fed83d3df033a4c648e3f9ac68afbd02f02fef845fc8251e"} Oct 02 16:59:10 crc kubenswrapper[4808]: I1002 16:59:10.200917 4808 generic.go:334] "Generic (PLEG): container finished" podID="5a2cb272-0bff-49dc-bf02-114dcfda4aea" containerID="fa1dcef324d54debc7e14b27ead8ec6020040e475f714145f1b3489b17524f76" exitCode=0 Oct 02 16:59:10 crc kubenswrapper[4808]: I1002 16:59:10.200964 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5e0c-account-create-75q46" event={"ID":"5a2cb272-0bff-49dc-bf02-114dcfda4aea","Type":"ContainerDied","Data":"fa1dcef324d54debc7e14b27ead8ec6020040e475f714145f1b3489b17524f76"} Oct 02 16:59:10 crc kubenswrapper[4808]: I1002 16:59:10.631087 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:59:11 crc kubenswrapper[4808]: I1002 16:59:11.706264 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4976-account-create-lskbv" Oct 02 16:59:11 crc kubenswrapper[4808]: I1002 16:59:11.727977 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5e0c-account-create-75q46" Oct 02 16:59:11 crc kubenswrapper[4808]: I1002 16:59:11.735562 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-731b-account-create-zrkfk" Oct 02 16:59:11 crc kubenswrapper[4808]: I1002 16:59:11.772315 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6szx\" (UniqueName: \"kubernetes.io/projected/e94ceeb9-6ff7-4fcd-8727-64bb034bef2f-kube-api-access-s6szx\") pod \"e94ceeb9-6ff7-4fcd-8727-64bb034bef2f\" (UID: \"e94ceeb9-6ff7-4fcd-8727-64bb034bef2f\") " Oct 02 16:59:11 crc kubenswrapper[4808]: I1002 16:59:11.772365 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqwg6\" (UniqueName: \"kubernetes.io/projected/b8311380-1e59-4d6c-b48f-c7dac382f325-kube-api-access-fqwg6\") pod \"b8311380-1e59-4d6c-b48f-c7dac382f325\" (UID: \"b8311380-1e59-4d6c-b48f-c7dac382f325\") " Oct 02 16:59:11 crc kubenswrapper[4808]: I1002 16:59:11.772452 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5fdj\" (UniqueName: \"kubernetes.io/projected/5a2cb272-0bff-49dc-bf02-114dcfda4aea-kube-api-access-n5fdj\") pod \"5a2cb272-0bff-49dc-bf02-114dcfda4aea\" (UID: \"5a2cb272-0bff-49dc-bf02-114dcfda4aea\") " Oct 02 16:59:11 crc kubenswrapper[4808]: I1002 16:59:11.778380 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e94ceeb9-6ff7-4fcd-8727-64bb034bef2f-kube-api-access-s6szx" (OuterVolumeSpecName: "kube-api-access-s6szx") pod "e94ceeb9-6ff7-4fcd-8727-64bb034bef2f" (UID: "e94ceeb9-6ff7-4fcd-8727-64bb034bef2f"). InnerVolumeSpecName "kube-api-access-s6szx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:59:11 crc kubenswrapper[4808]: I1002 16:59:11.780898 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8311380-1e59-4d6c-b48f-c7dac382f325-kube-api-access-fqwg6" (OuterVolumeSpecName: "kube-api-access-fqwg6") pod "b8311380-1e59-4d6c-b48f-c7dac382f325" (UID: "b8311380-1e59-4d6c-b48f-c7dac382f325"). InnerVolumeSpecName "kube-api-access-fqwg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:59:11 crc kubenswrapper[4808]: I1002 16:59:11.796666 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a2cb272-0bff-49dc-bf02-114dcfda4aea-kube-api-access-n5fdj" (OuterVolumeSpecName: "kube-api-access-n5fdj") pod "5a2cb272-0bff-49dc-bf02-114dcfda4aea" (UID: "5a2cb272-0bff-49dc-bf02-114dcfda4aea"). InnerVolumeSpecName "kube-api-access-n5fdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:59:11 crc kubenswrapper[4808]: I1002 16:59:11.874791 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6szx\" (UniqueName: \"kubernetes.io/projected/e94ceeb9-6ff7-4fcd-8727-64bb034bef2f-kube-api-access-s6szx\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:11 crc kubenswrapper[4808]: I1002 16:59:11.874838 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqwg6\" (UniqueName: \"kubernetes.io/projected/b8311380-1e59-4d6c-b48f-c7dac382f325-kube-api-access-fqwg6\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:11 crc kubenswrapper[4808]: I1002 16:59:11.874852 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5fdj\" (UniqueName: \"kubernetes.io/projected/5a2cb272-0bff-49dc-bf02-114dcfda4aea-kube-api-access-n5fdj\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:12 crc kubenswrapper[4808]: I1002 16:59:12.234836 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5e0c-account-create-75q46" event={"ID":"5a2cb272-0bff-49dc-bf02-114dcfda4aea","Type":"ContainerDied","Data":"a95fa1e3fe206618ea024cf0066feb2d34bdae8aea2b03cde79a6639e048d273"} Oct 02 16:59:12 crc kubenswrapper[4808]: I1002 16:59:12.234906 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a95fa1e3fe206618ea024cf0066feb2d34bdae8aea2b03cde79a6639e048d273" Oct 02 16:59:12 crc kubenswrapper[4808]: I1002 16:59:12.234999 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5e0c-account-create-75q46" Oct 02 16:59:12 crc kubenswrapper[4808]: I1002 16:59:12.239416 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4976-account-create-lskbv" Oct 02 16:59:12 crc kubenswrapper[4808]: I1002 16:59:12.239423 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4976-account-create-lskbv" event={"ID":"b8311380-1e59-4d6c-b48f-c7dac382f325","Type":"ContainerDied","Data":"99e3a017a6dd62320985729ff487d7788f9260550d2365a17b32d92c617ce048"} Oct 02 16:59:12 crc kubenswrapper[4808]: I1002 16:59:12.239479 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99e3a017a6dd62320985729ff487d7788f9260550d2365a17b32d92c617ce048" Oct 02 16:59:12 crc kubenswrapper[4808]: I1002 16:59:12.242121 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e31a6648-4159-4434-ac4f-7f5c142a650b","Type":"ContainerStarted","Data":"7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42"} Oct 02 16:59:12 crc kubenswrapper[4808]: I1002 16:59:12.242219 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e31a6648-4159-4434-ac4f-7f5c142a650b" containerName="ceilometer-central-agent" containerID="cri-o://2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e" gracePeriod=30 Oct 02 16:59:12 crc kubenswrapper[4808]: I1002 16:59:12.242280 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 16:59:12 crc kubenswrapper[4808]: I1002 16:59:12.242298 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e31a6648-4159-4434-ac4f-7f5c142a650b" containerName="proxy-httpd" containerID="cri-o://7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42" gracePeriod=30 Oct 02 16:59:12 crc kubenswrapper[4808]: I1002 16:59:12.242353 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e31a6648-4159-4434-ac4f-7f5c142a650b" containerName="sg-core" containerID="cri-o://d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9" gracePeriod=30 Oct 02 16:59:12 crc kubenswrapper[4808]: I1002 16:59:12.242409 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e31a6648-4159-4434-ac4f-7f5c142a650b" containerName="ceilometer-notification-agent" containerID="cri-o://4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e" gracePeriod=30 Oct 02 16:59:12 crc kubenswrapper[4808]: I1002 16:59:12.257675 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-731b-account-create-zrkfk" event={"ID":"e94ceeb9-6ff7-4fcd-8727-64bb034bef2f","Type":"ContainerDied","Data":"6633c2bcf65b623a175b29d52e1463fe4acd1311f7934d48ba3bb6b67ae4002d"} Oct 02 16:59:12 crc kubenswrapper[4808]: I1002 16:59:12.257705 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6633c2bcf65b623a175b29d52e1463fe4acd1311f7934d48ba3bb6b67ae4002d" Oct 02 16:59:12 crc kubenswrapper[4808]: I1002 16:59:12.257726 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-731b-account-create-zrkfk" Oct 02 16:59:12 crc kubenswrapper[4808]: I1002 16:59:12.267475 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.8861811149999999 podStartE2EDuration="6.267463172s" podCreationTimestamp="2025-10-02 16:59:06 +0000 UTC" firstStartedPulling="2025-10-02 16:59:07.075656931 +0000 UTC m=+1114.401185931" lastFinishedPulling="2025-10-02 16:59:11.456938988 +0000 UTC m=+1118.782467988" observedRunningTime="2025-10-02 16:59:12.262669272 +0000 UTC m=+1119.588198262" watchObservedRunningTime="2025-10-02 16:59:12.267463172 +0000 UTC m=+1119.592992172" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.078335 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.197365 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-combined-ca-bundle\") pod \"e31a6648-4159-4434-ac4f-7f5c142a650b\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.197423 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e31a6648-4159-4434-ac4f-7f5c142a650b-log-httpd\") pod \"e31a6648-4159-4434-ac4f-7f5c142a650b\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.197461 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e31a6648-4159-4434-ac4f-7f5c142a650b-run-httpd\") pod \"e31a6648-4159-4434-ac4f-7f5c142a650b\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.197546 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsckg\" (UniqueName: \"kubernetes.io/projected/e31a6648-4159-4434-ac4f-7f5c142a650b-kube-api-access-nsckg\") pod \"e31a6648-4159-4434-ac4f-7f5c142a650b\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.197630 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-scripts\") pod \"e31a6648-4159-4434-ac4f-7f5c142a650b\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.197658 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-sg-core-conf-yaml\") pod \"e31a6648-4159-4434-ac4f-7f5c142a650b\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.197738 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-config-data\") pod \"e31a6648-4159-4434-ac4f-7f5c142a650b\" (UID: \"e31a6648-4159-4434-ac4f-7f5c142a650b\") " Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.198553 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e31a6648-4159-4434-ac4f-7f5c142a650b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e31a6648-4159-4434-ac4f-7f5c142a650b" (UID: "e31a6648-4159-4434-ac4f-7f5c142a650b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.199372 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e31a6648-4159-4434-ac4f-7f5c142a650b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e31a6648-4159-4434-ac4f-7f5c142a650b" (UID: "e31a6648-4159-4434-ac4f-7f5c142a650b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.204359 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e31a6648-4159-4434-ac4f-7f5c142a650b-kube-api-access-nsckg" (OuterVolumeSpecName: "kube-api-access-nsckg") pod "e31a6648-4159-4434-ac4f-7f5c142a650b" (UID: "e31a6648-4159-4434-ac4f-7f5c142a650b"). InnerVolumeSpecName "kube-api-access-nsckg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.204500 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-scripts" (OuterVolumeSpecName: "scripts") pod "e31a6648-4159-4434-ac4f-7f5c142a650b" (UID: "e31a6648-4159-4434-ac4f-7f5c142a650b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.227821 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e31a6648-4159-4434-ac4f-7f5c142a650b" (UID: "e31a6648-4159-4434-ac4f-7f5c142a650b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.271506 4808 generic.go:334] "Generic (PLEG): container finished" podID="e31a6648-4159-4434-ac4f-7f5c142a650b" containerID="7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42" exitCode=0 Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.271536 4808 generic.go:334] "Generic (PLEG): container finished" podID="e31a6648-4159-4434-ac4f-7f5c142a650b" containerID="d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9" exitCode=2 Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.271547 4808 generic.go:334] "Generic (PLEG): container finished" podID="e31a6648-4159-4434-ac4f-7f5c142a650b" containerID="4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e" exitCode=0 Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.271556 4808 generic.go:334] "Generic (PLEG): container finished" podID="e31a6648-4159-4434-ac4f-7f5c142a650b" containerID="2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e" exitCode=0 Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.271565 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e31a6648-4159-4434-ac4f-7f5c142a650b","Type":"ContainerDied","Data":"7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42"} Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.271588 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.271632 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e31a6648-4159-4434-ac4f-7f5c142a650b","Type":"ContainerDied","Data":"d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9"} Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.271656 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e31a6648-4159-4434-ac4f-7f5c142a650b","Type":"ContainerDied","Data":"4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e"} Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.271676 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e31a6648-4159-4434-ac4f-7f5c142a650b","Type":"ContainerDied","Data":"2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e"} Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.271694 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e31a6648-4159-4434-ac4f-7f5c142a650b","Type":"ContainerDied","Data":"43e09574ae7a5a69885d462981ddb8b646c33bd95fbf171665817b3894a865e1"} Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.271708 4808 scope.go:117] "RemoveContainer" containerID="7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.293398 4808 scope.go:117] "RemoveContainer" containerID="d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.301419 4808 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e31a6648-4159-4434-ac4f-7f5c142a650b-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.301445 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsckg\" (UniqueName: \"kubernetes.io/projected/e31a6648-4159-4434-ac4f-7f5c142a650b-kube-api-access-nsckg\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.301454 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.301464 4808 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.301472 4808 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e31a6648-4159-4434-ac4f-7f5c142a650b-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.311226 4808 scope.go:117] "RemoveContainer" containerID="4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.313417 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e31a6648-4159-4434-ac4f-7f5c142a650b" (UID: "e31a6648-4159-4434-ac4f-7f5c142a650b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.314532 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-config-data" (OuterVolumeSpecName: "config-data") pod "e31a6648-4159-4434-ac4f-7f5c142a650b" (UID: "e31a6648-4159-4434-ac4f-7f5c142a650b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.338632 4808 scope.go:117] "RemoveContainer" containerID="2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.359309 4808 scope.go:117] "RemoveContainer" containerID="7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42" Oct 02 16:59:13 crc kubenswrapper[4808]: E1002 16:59:13.359780 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42\": container with ID starting with 7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42 not found: ID does not exist" containerID="7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.359811 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42"} err="failed to get container status \"7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42\": rpc error: code = NotFound desc = could not find container \"7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42\": container with ID starting with 7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42 not found: ID does not exist" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.359833 4808 scope.go:117] "RemoveContainer" containerID="d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9" Oct 02 16:59:13 crc kubenswrapper[4808]: E1002 16:59:13.360431 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9\": container with ID starting with d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9 not found: ID does not exist" containerID="d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.360453 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9"} err="failed to get container status \"d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9\": rpc error: code = NotFound desc = could not find container \"d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9\": container with ID starting with d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9 not found: ID does not exist" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.360467 4808 scope.go:117] "RemoveContainer" containerID="4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e" Oct 02 16:59:13 crc kubenswrapper[4808]: E1002 16:59:13.360907 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e\": container with ID starting with 4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e not found: ID does not exist" containerID="4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.360980 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e"} err="failed to get container status \"4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e\": rpc error: code = NotFound desc = could not find container \"4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e\": container with ID starting with 4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e not found: ID does not exist" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.361030 4808 scope.go:117] "RemoveContainer" containerID="2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e" Oct 02 16:59:13 crc kubenswrapper[4808]: E1002 16:59:13.361484 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e\": container with ID starting with 2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e not found: ID does not exist" containerID="2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.361512 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e"} err="failed to get container status \"2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e\": rpc error: code = NotFound desc = could not find container \"2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e\": container with ID starting with 2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e not found: ID does not exist" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.361528 4808 scope.go:117] "RemoveContainer" containerID="7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.361816 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42"} err="failed to get container status \"7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42\": rpc error: code = NotFound desc = could not find container \"7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42\": container with ID starting with 7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42 not found: ID does not exist" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.361837 4808 scope.go:117] "RemoveContainer" containerID="d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.362111 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9"} err="failed to get container status \"d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9\": rpc error: code = NotFound desc = could not find container \"d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9\": container with ID starting with d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9 not found: ID does not exist" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.362161 4808 scope.go:117] "RemoveContainer" containerID="4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.362538 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e"} err="failed to get container status \"4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e\": rpc error: code = NotFound desc = could not find container \"4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e\": container with ID starting with 4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e not found: ID does not exist" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.362558 4808 scope.go:117] "RemoveContainer" containerID="2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.362778 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e"} err="failed to get container status \"2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e\": rpc error: code = NotFound desc = could not find container \"2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e\": container with ID starting with 2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e not found: ID does not exist" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.362797 4808 scope.go:117] "RemoveContainer" containerID="7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.363024 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42"} err="failed to get container status \"7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42\": rpc error: code = NotFound desc = could not find container \"7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42\": container with ID starting with 7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42 not found: ID does not exist" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.363054 4808 scope.go:117] "RemoveContainer" containerID="d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.363338 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9"} err="failed to get container status \"d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9\": rpc error: code = NotFound desc = could not find container \"d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9\": container with ID starting with d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9 not found: ID does not exist" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.363367 4808 scope.go:117] "RemoveContainer" containerID="4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.363594 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e"} err="failed to get container status \"4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e\": rpc error: code = NotFound desc = could not find container \"4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e\": container with ID starting with 4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e not found: ID does not exist" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.363625 4808 scope.go:117] "RemoveContainer" containerID="2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.363866 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e"} err="failed to get container status \"2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e\": rpc error: code = NotFound desc = could not find container \"2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e\": container with ID starting with 2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e not found: ID does not exist" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.363885 4808 scope.go:117] "RemoveContainer" containerID="7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.364155 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42"} err="failed to get container status \"7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42\": rpc error: code = NotFound desc = could not find container \"7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42\": container with ID starting with 7fa8f8fc11271ebec18d60f9a45375314e044d9b0f3e70f109ea463eff2c6e42 not found: ID does not exist" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.364192 4808 scope.go:117] "RemoveContainer" containerID="d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.364510 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9"} err="failed to get container status \"d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9\": rpc error: code = NotFound desc = could not find container \"d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9\": container with ID starting with d57b23d607d432f3fba91462a47cb6cc694a13ab2a6eabf8c731ce4866a362e9 not found: ID does not exist" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.364558 4808 scope.go:117] "RemoveContainer" containerID="4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.364845 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e"} err="failed to get container status \"4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e\": rpc error: code = NotFound desc = could not find container \"4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e\": container with ID starting with 4255d57f58f2a66d6adb28d995bc5988e87b30179b69f2687398a7a1ed619a9e not found: ID does not exist" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.364874 4808 scope.go:117] "RemoveContainer" containerID="2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.366756 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e"} err="failed to get container status \"2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e\": rpc error: code = NotFound desc = could not find container \"2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e\": container with ID starting with 2b6abffa967987d3b43965fc817c4b48d1f3f48c138532714db7981ee4a8993e not found: ID does not exist" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.407622 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.407987 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e31a6648-4159-4434-ac4f-7f5c142a650b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.599411 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.610194 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.649947 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:59:13 crc kubenswrapper[4808]: E1002 16:59:13.650342 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31a6648-4159-4434-ac4f-7f5c142a650b" containerName="proxy-httpd" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.650359 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31a6648-4159-4434-ac4f-7f5c142a650b" containerName="proxy-httpd" Oct 02 16:59:13 crc kubenswrapper[4808]: E1002 16:59:13.650370 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31a6648-4159-4434-ac4f-7f5c142a650b" containerName="ceilometer-central-agent" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.650376 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31a6648-4159-4434-ac4f-7f5c142a650b" containerName="ceilometer-central-agent" Oct 02 16:59:13 crc kubenswrapper[4808]: E1002 16:59:13.650399 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8311380-1e59-4d6c-b48f-c7dac382f325" containerName="mariadb-account-create" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.650405 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8311380-1e59-4d6c-b48f-c7dac382f325" containerName="mariadb-account-create" Oct 02 16:59:13 crc kubenswrapper[4808]: E1002 16:59:13.650419 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31a6648-4159-4434-ac4f-7f5c142a650b" containerName="sg-core" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.650426 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31a6648-4159-4434-ac4f-7f5c142a650b" containerName="sg-core" Oct 02 16:59:13 crc kubenswrapper[4808]: E1002 16:59:13.650435 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a2cb272-0bff-49dc-bf02-114dcfda4aea" containerName="mariadb-account-create" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.650442 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a2cb272-0bff-49dc-bf02-114dcfda4aea" containerName="mariadb-account-create" Oct 02 16:59:13 crc kubenswrapper[4808]: E1002 16:59:13.650451 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31a6648-4159-4434-ac4f-7f5c142a650b" containerName="ceilometer-notification-agent" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.650457 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31a6648-4159-4434-ac4f-7f5c142a650b" containerName="ceilometer-notification-agent" Oct 02 16:59:13 crc kubenswrapper[4808]: E1002 16:59:13.650468 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e94ceeb9-6ff7-4fcd-8727-64bb034bef2f" containerName="mariadb-account-create" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.650474 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e94ceeb9-6ff7-4fcd-8727-64bb034bef2f" containerName="mariadb-account-create" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.650626 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e31a6648-4159-4434-ac4f-7f5c142a650b" containerName="ceilometer-central-agent" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.650636 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8311380-1e59-4d6c-b48f-c7dac382f325" containerName="mariadb-account-create" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.650652 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e31a6648-4159-4434-ac4f-7f5c142a650b" containerName="sg-core" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.650662 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a2cb272-0bff-49dc-bf02-114dcfda4aea" containerName="mariadb-account-create" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.650670 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e31a6648-4159-4434-ac4f-7f5c142a650b" containerName="proxy-httpd" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.650681 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e31a6648-4159-4434-ac4f-7f5c142a650b" containerName="ceilometer-notification-agent" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.650689 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e94ceeb9-6ff7-4fcd-8727-64bb034bef2f" containerName="mariadb-account-create" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.652066 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.665665 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.665873 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.672282 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.746792 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hmw58"] Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.747819 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hmw58" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.751253 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.751457 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.755777 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-wswpj" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.764109 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hmw58"] Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.822132 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-scripts\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.822318 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-run-httpd\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.822417 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.822519 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-log-httpd\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.822567 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8kmr\" (UniqueName: \"kubernetes.io/projected/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-kube-api-access-w8kmr\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.822585 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-config-data\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.822602 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.923614 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6x8d\" (UniqueName: \"kubernetes.io/projected/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-kube-api-access-h6x8d\") pod \"nova-cell0-conductor-db-sync-hmw58\" (UID: \"81cc1bfd-c3b7-49e1-9bc7-466483bde77b\") " pod="openstack/nova-cell0-conductor-db-sync-hmw58" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.923670 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.923714 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-log-httpd\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.923734 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-config-data\") pod \"nova-cell0-conductor-db-sync-hmw58\" (UID: \"81cc1bfd-c3b7-49e1-9bc7-466483bde77b\") " pod="openstack/nova-cell0-conductor-db-sync-hmw58" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.923759 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-hmw58\" (UID: \"81cc1bfd-c3b7-49e1-9bc7-466483bde77b\") " pod="openstack/nova-cell0-conductor-db-sync-hmw58" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.923784 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8kmr\" (UniqueName: \"kubernetes.io/projected/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-kube-api-access-w8kmr\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.923804 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-config-data\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.923818 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.923852 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-scripts\") pod \"nova-cell0-conductor-db-sync-hmw58\" (UID: \"81cc1bfd-c3b7-49e1-9bc7-466483bde77b\") " pod="openstack/nova-cell0-conductor-db-sync-hmw58" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.923890 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-scripts\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.923919 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-run-httpd\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.924795 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-run-httpd\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.924882 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-log-httpd\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.928942 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.929652 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-scripts\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.932796 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-config-data\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.945837 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8kmr\" (UniqueName: \"kubernetes.io/projected/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-kube-api-access-w8kmr\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.947620 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " pod="openstack/ceilometer-0" Oct 02 16:59:13 crc kubenswrapper[4808]: I1002 16:59:13.975343 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:59:14 crc kubenswrapper[4808]: I1002 16:59:14.025079 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6x8d\" (UniqueName: \"kubernetes.io/projected/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-kube-api-access-h6x8d\") pod \"nova-cell0-conductor-db-sync-hmw58\" (UID: \"81cc1bfd-c3b7-49e1-9bc7-466483bde77b\") " pod="openstack/nova-cell0-conductor-db-sync-hmw58" Oct 02 16:59:14 crc kubenswrapper[4808]: I1002 16:59:14.025172 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-config-data\") pod \"nova-cell0-conductor-db-sync-hmw58\" (UID: \"81cc1bfd-c3b7-49e1-9bc7-466483bde77b\") " pod="openstack/nova-cell0-conductor-db-sync-hmw58" Oct 02 16:59:14 crc kubenswrapper[4808]: I1002 16:59:14.025197 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-hmw58\" (UID: \"81cc1bfd-c3b7-49e1-9bc7-466483bde77b\") " pod="openstack/nova-cell0-conductor-db-sync-hmw58" Oct 02 16:59:14 crc kubenswrapper[4808]: I1002 16:59:14.025242 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-scripts\") pod \"nova-cell0-conductor-db-sync-hmw58\" (UID: \"81cc1bfd-c3b7-49e1-9bc7-466483bde77b\") " pod="openstack/nova-cell0-conductor-db-sync-hmw58" Oct 02 16:59:14 crc kubenswrapper[4808]: I1002 16:59:14.034326 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-hmw58\" (UID: \"81cc1bfd-c3b7-49e1-9bc7-466483bde77b\") " pod="openstack/nova-cell0-conductor-db-sync-hmw58" Oct 02 16:59:14 crc kubenswrapper[4808]: I1002 16:59:14.034426 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-scripts\") pod \"nova-cell0-conductor-db-sync-hmw58\" (UID: \"81cc1bfd-c3b7-49e1-9bc7-466483bde77b\") " pod="openstack/nova-cell0-conductor-db-sync-hmw58" Oct 02 16:59:14 crc kubenswrapper[4808]: I1002 16:59:14.034782 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-config-data\") pod \"nova-cell0-conductor-db-sync-hmw58\" (UID: \"81cc1bfd-c3b7-49e1-9bc7-466483bde77b\") " pod="openstack/nova-cell0-conductor-db-sync-hmw58" Oct 02 16:59:14 crc kubenswrapper[4808]: I1002 16:59:14.060057 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6x8d\" (UniqueName: \"kubernetes.io/projected/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-kube-api-access-h6x8d\") pod \"nova-cell0-conductor-db-sync-hmw58\" (UID: \"81cc1bfd-c3b7-49e1-9bc7-466483bde77b\") " pod="openstack/nova-cell0-conductor-db-sync-hmw58" Oct 02 16:59:14 crc kubenswrapper[4808]: I1002 16:59:14.061648 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hmw58" Oct 02 16:59:14 crc kubenswrapper[4808]: I1002 16:59:14.314841 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hmw58"] Oct 02 16:59:14 crc kubenswrapper[4808]: W1002 16:59:14.325764 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81cc1bfd_c3b7_49e1_9bc7_466483bde77b.slice/crio-79f029c7bea93f73642f9464f6f75e853e05b4faac48d04ad9a84d82ec94f89f WatchSource:0}: Error finding container 79f029c7bea93f73642f9464f6f75e853e05b4faac48d04ad9a84d82ec94f89f: Status 404 returned error can't find the container with id 79f029c7bea93f73642f9464f6f75e853e05b4faac48d04ad9a84d82ec94f89f Oct 02 16:59:14 crc kubenswrapper[4808]: I1002 16:59:14.408455 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:59:14 crc kubenswrapper[4808]: W1002 16:59:14.413446 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod596a4d5a_b4ba_466b_a9e5_ac4c853efcb7.slice/crio-cce32853ea1e26a3b1a35a26b1e8de066a35b21f3d7ac227030f2ad07a363b55 WatchSource:0}: Error finding container cce32853ea1e26a3b1a35a26b1e8de066a35b21f3d7ac227030f2ad07a363b55: Status 404 returned error can't find the container with id cce32853ea1e26a3b1a35a26b1e8de066a35b21f3d7ac227030f2ad07a363b55 Oct 02 16:59:15 crc kubenswrapper[4808]: I1002 16:59:15.300702 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7","Type":"ContainerStarted","Data":"80ed9069762936ea68104c7c75c37bfd9b082fa028cd3d54e34778db6bf53a6a"} Oct 02 16:59:15 crc kubenswrapper[4808]: I1002 16:59:15.300961 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7","Type":"ContainerStarted","Data":"cce32853ea1e26a3b1a35a26b1e8de066a35b21f3d7ac227030f2ad07a363b55"} Oct 02 16:59:15 crc kubenswrapper[4808]: I1002 16:59:15.302228 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hmw58" event={"ID":"81cc1bfd-c3b7-49e1-9bc7-466483bde77b","Type":"ContainerStarted","Data":"79f029c7bea93f73642f9464f6f75e853e05b4faac48d04ad9a84d82ec94f89f"} Oct 02 16:59:15 crc kubenswrapper[4808]: I1002 16:59:15.406670 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e31a6648-4159-4434-ac4f-7f5c142a650b" path="/var/lib/kubelet/pods/e31a6648-4159-4434-ac4f-7f5c142a650b/volumes" Oct 02 16:59:16 crc kubenswrapper[4808]: I1002 16:59:16.317387 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7","Type":"ContainerStarted","Data":"3d0dd557125f47597903502f7aea47322f8712e52bf3c8140a68ed8715fe94da"} Oct 02 16:59:17 crc kubenswrapper[4808]: I1002 16:59:17.329429 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7","Type":"ContainerStarted","Data":"94522ef9e646b66e17dc4bfa2f54fe131a24052e954ad1d679f5b93845992bf0"} Oct 02 16:59:21 crc kubenswrapper[4808]: I1002 16:59:21.368391 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hmw58" event={"ID":"81cc1bfd-c3b7-49e1-9bc7-466483bde77b","Type":"ContainerStarted","Data":"af9a06024f8a93be5e96603b479ae8bae2cafda0340872e07d2f8d59b2768dbd"} Oct 02 16:59:21 crc kubenswrapper[4808]: I1002 16:59:21.373120 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7","Type":"ContainerStarted","Data":"fe7ced19e6b556a1b7a881eaf296ab31a77bc3a70e065fde7fd3dd34a1714fe9"} Oct 02 16:59:21 crc kubenswrapper[4808]: I1002 16:59:21.373339 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 16:59:21 crc kubenswrapper[4808]: I1002 16:59:21.399316 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-hmw58" podStartSLOduration=1.947090957 podStartE2EDuration="8.399283513s" podCreationTimestamp="2025-10-02 16:59:13 +0000 UTC" firstStartedPulling="2025-10-02 16:59:14.327943778 +0000 UTC m=+1121.653472778" lastFinishedPulling="2025-10-02 16:59:20.780136324 +0000 UTC m=+1128.105665334" observedRunningTime="2025-10-02 16:59:21.389897258 +0000 UTC m=+1128.715426318" watchObservedRunningTime="2025-10-02 16:59:21.399283513 +0000 UTC m=+1128.724812553" Oct 02 16:59:21 crc kubenswrapper[4808]: I1002 16:59:21.440148 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.078719979 podStartE2EDuration="8.440123751s" podCreationTimestamp="2025-10-02 16:59:13 +0000 UTC" firstStartedPulling="2025-10-02 16:59:14.415899866 +0000 UTC m=+1121.741428866" lastFinishedPulling="2025-10-02 16:59:20.777303638 +0000 UTC m=+1128.102832638" observedRunningTime="2025-10-02 16:59:21.422300607 +0000 UTC m=+1128.747829647" watchObservedRunningTime="2025-10-02 16:59:21.440123751 +0000 UTC m=+1128.765652791" Oct 02 16:59:30 crc kubenswrapper[4808]: I1002 16:59:30.480746 4808 generic.go:334] "Generic (PLEG): container finished" podID="81cc1bfd-c3b7-49e1-9bc7-466483bde77b" containerID="af9a06024f8a93be5e96603b479ae8bae2cafda0340872e07d2f8d59b2768dbd" exitCode=0 Oct 02 16:59:30 crc kubenswrapper[4808]: I1002 16:59:30.480852 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hmw58" event={"ID":"81cc1bfd-c3b7-49e1-9bc7-466483bde77b","Type":"ContainerDied","Data":"af9a06024f8a93be5e96603b479ae8bae2cafda0340872e07d2f8d59b2768dbd"} Oct 02 16:59:31 crc kubenswrapper[4808]: I1002 16:59:31.928029 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hmw58" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.060628 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-config-data\") pod \"81cc1bfd-c3b7-49e1-9bc7-466483bde77b\" (UID: \"81cc1bfd-c3b7-49e1-9bc7-466483bde77b\") " Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.060814 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-combined-ca-bundle\") pod \"81cc1bfd-c3b7-49e1-9bc7-466483bde77b\" (UID: \"81cc1bfd-c3b7-49e1-9bc7-466483bde77b\") " Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.060885 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6x8d\" (UniqueName: \"kubernetes.io/projected/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-kube-api-access-h6x8d\") pod \"81cc1bfd-c3b7-49e1-9bc7-466483bde77b\" (UID: \"81cc1bfd-c3b7-49e1-9bc7-466483bde77b\") " Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.060972 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-scripts\") pod \"81cc1bfd-c3b7-49e1-9bc7-466483bde77b\" (UID: \"81cc1bfd-c3b7-49e1-9bc7-466483bde77b\") " Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.065975 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-kube-api-access-h6x8d" (OuterVolumeSpecName: "kube-api-access-h6x8d") pod "81cc1bfd-c3b7-49e1-9bc7-466483bde77b" (UID: "81cc1bfd-c3b7-49e1-9bc7-466483bde77b"). InnerVolumeSpecName "kube-api-access-h6x8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.068052 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-scripts" (OuterVolumeSpecName: "scripts") pod "81cc1bfd-c3b7-49e1-9bc7-466483bde77b" (UID: "81cc1bfd-c3b7-49e1-9bc7-466483bde77b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.102453 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81cc1bfd-c3b7-49e1-9bc7-466483bde77b" (UID: "81cc1bfd-c3b7-49e1-9bc7-466483bde77b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.105202 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-config-data" (OuterVolumeSpecName: "config-data") pod "81cc1bfd-c3b7-49e1-9bc7-466483bde77b" (UID: "81cc1bfd-c3b7-49e1-9bc7-466483bde77b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.163958 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.164023 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6x8d\" (UniqueName: \"kubernetes.io/projected/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-kube-api-access-h6x8d\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.164051 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.164073 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81cc1bfd-c3b7-49e1-9bc7-466483bde77b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.502270 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hmw58" event={"ID":"81cc1bfd-c3b7-49e1-9bc7-466483bde77b","Type":"ContainerDied","Data":"79f029c7bea93f73642f9464f6f75e853e05b4faac48d04ad9a84d82ec94f89f"} Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.502310 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79f029c7bea93f73642f9464f6f75e853e05b4faac48d04ad9a84d82ec94f89f" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.502361 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hmw58" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.595638 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 16:59:32 crc kubenswrapper[4808]: E1002 16:59:32.596337 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81cc1bfd-c3b7-49e1-9bc7-466483bde77b" containerName="nova-cell0-conductor-db-sync" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.596356 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="81cc1bfd-c3b7-49e1-9bc7-466483bde77b" containerName="nova-cell0-conductor-db-sync" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.596589 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="81cc1bfd-c3b7-49e1-9bc7-466483bde77b" containerName="nova-cell0-conductor-db-sync" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.597372 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.601340 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.601486 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-wswpj" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.609885 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.639336 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxt2h\" (UniqueName: \"kubernetes.io/projected/eb046ea2-150a-494e-bfa5-5b4283f62456-kube-api-access-nxt2h\") pod \"nova-cell0-conductor-0\" (UID: \"eb046ea2-150a-494e-bfa5-5b4283f62456\") " pod="openstack/nova-cell0-conductor-0" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.639450 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb046ea2-150a-494e-bfa5-5b4283f62456-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"eb046ea2-150a-494e-bfa5-5b4283f62456\") " pod="openstack/nova-cell0-conductor-0" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.639526 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb046ea2-150a-494e-bfa5-5b4283f62456-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"eb046ea2-150a-494e-bfa5-5b4283f62456\") " pod="openstack/nova-cell0-conductor-0" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.740905 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb046ea2-150a-494e-bfa5-5b4283f62456-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"eb046ea2-150a-494e-bfa5-5b4283f62456\") " pod="openstack/nova-cell0-conductor-0" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.741047 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxt2h\" (UniqueName: \"kubernetes.io/projected/eb046ea2-150a-494e-bfa5-5b4283f62456-kube-api-access-nxt2h\") pod \"nova-cell0-conductor-0\" (UID: \"eb046ea2-150a-494e-bfa5-5b4283f62456\") " pod="openstack/nova-cell0-conductor-0" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.741083 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb046ea2-150a-494e-bfa5-5b4283f62456-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"eb046ea2-150a-494e-bfa5-5b4283f62456\") " pod="openstack/nova-cell0-conductor-0" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.747547 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb046ea2-150a-494e-bfa5-5b4283f62456-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"eb046ea2-150a-494e-bfa5-5b4283f62456\") " pod="openstack/nova-cell0-conductor-0" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.748896 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb046ea2-150a-494e-bfa5-5b4283f62456-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"eb046ea2-150a-494e-bfa5-5b4283f62456\") " pod="openstack/nova-cell0-conductor-0" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.760013 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxt2h\" (UniqueName: \"kubernetes.io/projected/eb046ea2-150a-494e-bfa5-5b4283f62456-kube-api-access-nxt2h\") pod \"nova-cell0-conductor-0\" (UID: \"eb046ea2-150a-494e-bfa5-5b4283f62456\") " pod="openstack/nova-cell0-conductor-0" Oct 02 16:59:32 crc kubenswrapper[4808]: I1002 16:59:32.921857 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 16:59:33 crc kubenswrapper[4808]: I1002 16:59:33.453718 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 16:59:33 crc kubenswrapper[4808]: I1002 16:59:33.515120 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"eb046ea2-150a-494e-bfa5-5b4283f62456","Type":"ContainerStarted","Data":"943015aabea74d2551a92331b78710e7d763d4578d3b0abb41ae2887d8ad658b"} Oct 02 16:59:34 crc kubenswrapper[4808]: I1002 16:59:34.530273 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"eb046ea2-150a-494e-bfa5-5b4283f62456","Type":"ContainerStarted","Data":"600e11c6d97ff7aaa49fb263f5a727b85d53b9e88201ff91ca8568c78ac40fbb"} Oct 02 16:59:34 crc kubenswrapper[4808]: I1002 16:59:34.530727 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 02 16:59:34 crc kubenswrapper[4808]: I1002 16:59:34.560176 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.560147999 podStartE2EDuration="2.560147999s" podCreationTimestamp="2025-10-02 16:59:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:59:34.550021874 +0000 UTC m=+1141.875550914" watchObservedRunningTime="2025-10-02 16:59:34.560147999 +0000 UTC m=+1141.885677029" Oct 02 16:59:42 crc kubenswrapper[4808]: I1002 16:59:42.639596 4808 generic.go:334] "Generic (PLEG): container finished" podID="818e02a3-653a-4b36-a68b-9d4e2df0b71e" containerID="a918ee069243e5dc0b7673d53697e23876162204008e5ae0c51065fa9012a203" exitCode=1 Oct 02 16:59:42 crc kubenswrapper[4808]: I1002 16:59:42.639718 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" event={"ID":"818e02a3-653a-4b36-a68b-9d4e2df0b71e","Type":"ContainerDied","Data":"a918ee069243e5dc0b7673d53697e23876162204008e5ae0c51065fa9012a203"} Oct 02 16:59:42 crc kubenswrapper[4808]: I1002 16:59:42.640491 4808 scope.go:117] "RemoveContainer" containerID="72dfceead77ed393455b2e7fbcb7b164af0fc5382ab0ba62ad1f243b06b38645" Oct 02 16:59:42 crc kubenswrapper[4808]: I1002 16:59:42.641381 4808 scope.go:117] "RemoveContainer" containerID="a918ee069243e5dc0b7673d53697e23876162204008e5ae0c51065fa9012a203" Oct 02 16:59:42 crc kubenswrapper[4808]: E1002 16:59:42.641811 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=keystone-operator-controller-manager-65f4bcbb94-svkdd_openstack-operators(818e02a3-653a-4b36-a68b-9d4e2df0b71e)\"" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" podUID="818e02a3-653a-4b36-a68b-9d4e2df0b71e" Oct 02 16:59:42 crc kubenswrapper[4808]: I1002 16:59:42.969227 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.467701 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-x8vmm"] Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.468782 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x8vmm" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.471051 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.480071 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-x8vmm"] Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.480711 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.656555 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x8vmm\" (UID: \"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958\") " pod="openstack/nova-cell0-cell-mapping-x8vmm" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.656599 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-scripts\") pod \"nova-cell0-cell-mapping-x8vmm\" (UID: \"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958\") " pod="openstack/nova-cell0-cell-mapping-x8vmm" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.656624 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-config-data\") pod \"nova-cell0-cell-mapping-x8vmm\" (UID: \"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958\") " pod="openstack/nova-cell0-cell-mapping-x8vmm" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.656662 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8hfg\" (UniqueName: \"kubernetes.io/projected/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-kube-api-access-r8hfg\") pod \"nova-cell0-cell-mapping-x8vmm\" (UID: \"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958\") " pod="openstack/nova-cell0-cell-mapping-x8vmm" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.670623 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.672939 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.674937 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.704860 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.758619 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x8vmm\" (UID: \"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958\") " pod="openstack/nova-cell0-cell-mapping-x8vmm" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.758674 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-scripts\") pod \"nova-cell0-cell-mapping-x8vmm\" (UID: \"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958\") " pod="openstack/nova-cell0-cell-mapping-x8vmm" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.758704 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-config-data\") pod \"nova-cell0-cell-mapping-x8vmm\" (UID: \"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958\") " pod="openstack/nova-cell0-cell-mapping-x8vmm" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.758748 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8hfg\" (UniqueName: \"kubernetes.io/projected/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-kube-api-access-r8hfg\") pod \"nova-cell0-cell-mapping-x8vmm\" (UID: \"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958\") " pod="openstack/nova-cell0-cell-mapping-x8vmm" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.766029 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-scripts\") pod \"nova-cell0-cell-mapping-x8vmm\" (UID: \"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958\") " pod="openstack/nova-cell0-cell-mapping-x8vmm" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.768625 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-config-data\") pod \"nova-cell0-cell-mapping-x8vmm\" (UID: \"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958\") " pod="openstack/nova-cell0-cell-mapping-x8vmm" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.786524 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x8vmm\" (UID: \"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958\") " pod="openstack/nova-cell0-cell-mapping-x8vmm" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.789481 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.790834 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.795962 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.796345 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8hfg\" (UniqueName: \"kubernetes.io/projected/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-kube-api-access-r8hfg\") pod \"nova-cell0-cell-mapping-x8vmm\" (UID: \"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958\") " pod="openstack/nova-cell0-cell-mapping-x8vmm" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.817732 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.869206 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4b16f51-ca0e-41eb-9033-86a5f7ee6091-config-data\") pod \"nova-scheduler-0\" (UID: \"b4b16f51-ca0e-41eb-9033-86a5f7ee6091\") " pod="openstack/nova-scheduler-0" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.869407 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4b16f51-ca0e-41eb-9033-86a5f7ee6091-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b4b16f51-ca0e-41eb-9033-86a5f7ee6091\") " pod="openstack/nova-scheduler-0" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.869429 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rkvc\" (UniqueName: \"kubernetes.io/projected/b4b16f51-ca0e-41eb-9033-86a5f7ee6091-kube-api-access-4rkvc\") pod \"nova-scheduler-0\" (UID: \"b4b16f51-ca0e-41eb-9033-86a5f7ee6091\") " pod="openstack/nova-scheduler-0" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.869740 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.881047 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.883270 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.900957 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.952068 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-85kbf"] Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.957549 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.970737 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4b16f51-ca0e-41eb-9033-86a5f7ee6091-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b4b16f51-ca0e-41eb-9033-86a5f7ee6091\") " pod="openstack/nova-scheduler-0" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.971155 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rkvc\" (UniqueName: \"kubernetes.io/projected/b4b16f51-ca0e-41eb-9033-86a5f7ee6091-kube-api-access-4rkvc\") pod \"nova-scheduler-0\" (UID: \"b4b16f51-ca0e-41eb-9033-86a5f7ee6091\") " pod="openstack/nova-scheduler-0" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.971280 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7d214c5-98b6-46db-b77a-5415a89227d0-logs\") pod \"nova-metadata-0\" (UID: \"f7d214c5-98b6-46db-b77a-5415a89227d0\") " pod="openstack/nova-metadata-0" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.971399 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7d214c5-98b6-46db-b77a-5415a89227d0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f7d214c5-98b6-46db-b77a-5415a89227d0\") " pod="openstack/nova-metadata-0" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.971479 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7d214c5-98b6-46db-b77a-5415a89227d0-config-data\") pod \"nova-metadata-0\" (UID: \"f7d214c5-98b6-46db-b77a-5415a89227d0\") " pod="openstack/nova-metadata-0" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.971545 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndgp2\" (UniqueName: \"kubernetes.io/projected/f7d214c5-98b6-46db-b77a-5415a89227d0-kube-api-access-ndgp2\") pod \"nova-metadata-0\" (UID: \"f7d214c5-98b6-46db-b77a-5415a89227d0\") " pod="openstack/nova-metadata-0" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.971690 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4b16f51-ca0e-41eb-9033-86a5f7ee6091-config-data\") pod \"nova-scheduler-0\" (UID: \"b4b16f51-ca0e-41eb-9033-86a5f7ee6091\") " pod="openstack/nova-scheduler-0" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.977310 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4b16f51-ca0e-41eb-9033-86a5f7ee6091-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b4b16f51-ca0e-41eb-9033-86a5f7ee6091\") " pod="openstack/nova-scheduler-0" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.979995 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-85kbf"] Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.980210 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4b16f51-ca0e-41eb-9033-86a5f7ee6091-config-data\") pod \"nova-scheduler-0\" (UID: \"b4b16f51-ca0e-41eb-9033-86a5f7ee6091\") " pod="openstack/nova-scheduler-0" Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.997313 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 16:59:43 crc kubenswrapper[4808]: I1002 16:59:43.998367 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.006522 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.007513 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.007667 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.016880 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rkvc\" (UniqueName: \"kubernetes.io/projected/b4b16f51-ca0e-41eb-9033-86a5f7ee6091-kube-api-access-4rkvc\") pod \"nova-scheduler-0\" (UID: \"b4b16f51-ca0e-41eb-9033-86a5f7ee6091\") " pod="openstack/nova-scheduler-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.073587 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndgp2\" (UniqueName: \"kubernetes.io/projected/f7d214c5-98b6-46db-b77a-5415a89227d0-kube-api-access-ndgp2\") pod \"nova-metadata-0\" (UID: \"f7d214c5-98b6-46db-b77a-5415a89227d0\") " pod="openstack/nova-metadata-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.073638 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-ovsdbserver-nb\") pod \"dnsmasq-dns-566b5b7845-85kbf\" (UID: \"fe4d6d85-749d-4ca8-a962-e49678672538\") " pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.073673 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c081cab-30c8-4952-8527-65355562b158-logs\") pod \"nova-api-0\" (UID: \"6c081cab-30c8-4952-8527-65355562b158\") " pod="openstack/nova-api-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.073696 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-config\") pod \"dnsmasq-dns-566b5b7845-85kbf\" (UID: \"fe4d6d85-749d-4ca8-a962-e49678672538\") " pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.073724 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c081cab-30c8-4952-8527-65355562b158-config-data\") pod \"nova-api-0\" (UID: \"6c081cab-30c8-4952-8527-65355562b158\") " pod="openstack/nova-api-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.073750 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dmm8\" (UniqueName: \"kubernetes.io/projected/fe4d6d85-749d-4ca8-a962-e49678672538-kube-api-access-8dmm8\") pod \"dnsmasq-dns-566b5b7845-85kbf\" (UID: \"fe4d6d85-749d-4ca8-a962-e49678672538\") " pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.073850 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-ovsdbserver-sb\") pod \"dnsmasq-dns-566b5b7845-85kbf\" (UID: \"fe4d6d85-749d-4ca8-a962-e49678672538\") " pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.073883 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt682\" (UniqueName: \"kubernetes.io/projected/6c081cab-30c8-4952-8527-65355562b158-kube-api-access-jt682\") pod \"nova-api-0\" (UID: \"6c081cab-30c8-4952-8527-65355562b158\") " pod="openstack/nova-api-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.073900 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7d214c5-98b6-46db-b77a-5415a89227d0-logs\") pod \"nova-metadata-0\" (UID: \"f7d214c5-98b6-46db-b77a-5415a89227d0\") " pod="openstack/nova-metadata-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.073966 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-dns-svc\") pod \"dnsmasq-dns-566b5b7845-85kbf\" (UID: \"fe4d6d85-749d-4ca8-a962-e49678672538\") " pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.074014 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c081cab-30c8-4952-8527-65355562b158-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6c081cab-30c8-4952-8527-65355562b158\") " pod="openstack/nova-api-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.074036 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7d214c5-98b6-46db-b77a-5415a89227d0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f7d214c5-98b6-46db-b77a-5415a89227d0\") " pod="openstack/nova-metadata-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.074092 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7d214c5-98b6-46db-b77a-5415a89227d0-config-data\") pod \"nova-metadata-0\" (UID: \"f7d214c5-98b6-46db-b77a-5415a89227d0\") " pod="openstack/nova-metadata-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.074211 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7d214c5-98b6-46db-b77a-5415a89227d0-logs\") pod \"nova-metadata-0\" (UID: \"f7d214c5-98b6-46db-b77a-5415a89227d0\") " pod="openstack/nova-metadata-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.077758 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7d214c5-98b6-46db-b77a-5415a89227d0-config-data\") pod \"nova-metadata-0\" (UID: \"f7d214c5-98b6-46db-b77a-5415a89227d0\") " pod="openstack/nova-metadata-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.090186 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7d214c5-98b6-46db-b77a-5415a89227d0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f7d214c5-98b6-46db-b77a-5415a89227d0\") " pod="openstack/nova-metadata-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.092124 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x8vmm" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.094135 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndgp2\" (UniqueName: \"kubernetes.io/projected/f7d214c5-98b6-46db-b77a-5415a89227d0-kube-api-access-ndgp2\") pod \"nova-metadata-0\" (UID: \"f7d214c5-98b6-46db-b77a-5415a89227d0\") " pod="openstack/nova-metadata-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.175732 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c081cab-30c8-4952-8527-65355562b158-config-data\") pod \"nova-api-0\" (UID: \"6c081cab-30c8-4952-8527-65355562b158\") " pod="openstack/nova-api-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.175804 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dmm8\" (UniqueName: \"kubernetes.io/projected/fe4d6d85-749d-4ca8-a962-e49678672538-kube-api-access-8dmm8\") pod \"dnsmasq-dns-566b5b7845-85kbf\" (UID: \"fe4d6d85-749d-4ca8-a962-e49678672538\") " pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.175850 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cd4a7e0-6aa4-4418-889a-2112878249bd-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7cd4a7e0-6aa4-4418-889a-2112878249bd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.175911 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-ovsdbserver-sb\") pod \"dnsmasq-dns-566b5b7845-85kbf\" (UID: \"fe4d6d85-749d-4ca8-a962-e49678672538\") " pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.175962 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt682\" (UniqueName: \"kubernetes.io/projected/6c081cab-30c8-4952-8527-65355562b158-kube-api-access-jt682\") pod \"nova-api-0\" (UID: \"6c081cab-30c8-4952-8527-65355562b158\") " pod="openstack/nova-api-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.175989 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-dns-svc\") pod \"dnsmasq-dns-566b5b7845-85kbf\" (UID: \"fe4d6d85-749d-4ca8-a962-e49678672538\") " pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.176068 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c081cab-30c8-4952-8527-65355562b158-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6c081cab-30c8-4952-8527-65355562b158\") " pod="openstack/nova-api-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.176570 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2jb2\" (UniqueName: \"kubernetes.io/projected/7cd4a7e0-6aa4-4418-889a-2112878249bd-kube-api-access-l2jb2\") pod \"nova-cell1-novncproxy-0\" (UID: \"7cd4a7e0-6aa4-4418-889a-2112878249bd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.177717 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-ovsdbserver-nb\") pod \"dnsmasq-dns-566b5b7845-85kbf\" (UID: \"fe4d6d85-749d-4ca8-a962-e49678672538\") " pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.177821 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c081cab-30c8-4952-8527-65355562b158-logs\") pod \"nova-api-0\" (UID: \"6c081cab-30c8-4952-8527-65355562b158\") " pod="openstack/nova-api-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.177881 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-config\") pod \"dnsmasq-dns-566b5b7845-85kbf\" (UID: \"fe4d6d85-749d-4ca8-a962-e49678672538\") " pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.177943 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cd4a7e0-6aa4-4418-889a-2112878249bd-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7cd4a7e0-6aa4-4418-889a-2112878249bd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.178279 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-ovsdbserver-sb\") pod \"dnsmasq-dns-566b5b7845-85kbf\" (UID: \"fe4d6d85-749d-4ca8-a962-e49678672538\") " pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.178662 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-dns-svc\") pod \"dnsmasq-dns-566b5b7845-85kbf\" (UID: \"fe4d6d85-749d-4ca8-a962-e49678672538\") " pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.178796 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c081cab-30c8-4952-8527-65355562b158-logs\") pod \"nova-api-0\" (UID: \"6c081cab-30c8-4952-8527-65355562b158\") " pod="openstack/nova-api-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.178810 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-ovsdbserver-nb\") pod \"dnsmasq-dns-566b5b7845-85kbf\" (UID: \"fe4d6d85-749d-4ca8-a962-e49678672538\") " pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.179407 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-config\") pod \"dnsmasq-dns-566b5b7845-85kbf\" (UID: \"fe4d6d85-749d-4ca8-a962-e49678672538\") " pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.182666 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c081cab-30c8-4952-8527-65355562b158-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6c081cab-30c8-4952-8527-65355562b158\") " pod="openstack/nova-api-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.186099 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c081cab-30c8-4952-8527-65355562b158-config-data\") pod \"nova-api-0\" (UID: \"6c081cab-30c8-4952-8527-65355562b158\") " pod="openstack/nova-api-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.203670 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dmm8\" (UniqueName: \"kubernetes.io/projected/fe4d6d85-749d-4ca8-a962-e49678672538-kube-api-access-8dmm8\") pod \"dnsmasq-dns-566b5b7845-85kbf\" (UID: \"fe4d6d85-749d-4ca8-a962-e49678672538\") " pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.213795 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt682\" (UniqueName: \"kubernetes.io/projected/6c081cab-30c8-4952-8527-65355562b158-kube-api-access-jt682\") pod \"nova-api-0\" (UID: \"6c081cab-30c8-4952-8527-65355562b158\") " pod="openstack/nova-api-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.219261 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.268222 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.282523 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2jb2\" (UniqueName: \"kubernetes.io/projected/7cd4a7e0-6aa4-4418-889a-2112878249bd-kube-api-access-l2jb2\") pod \"nova-cell1-novncproxy-0\" (UID: \"7cd4a7e0-6aa4-4418-889a-2112878249bd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.282656 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cd4a7e0-6aa4-4418-889a-2112878249bd-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7cd4a7e0-6aa4-4418-889a-2112878249bd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.282736 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cd4a7e0-6aa4-4418-889a-2112878249bd-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7cd4a7e0-6aa4-4418-889a-2112878249bd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.290911 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cd4a7e0-6aa4-4418-889a-2112878249bd-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7cd4a7e0-6aa4-4418-889a-2112878249bd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.291040 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cd4a7e0-6aa4-4418-889a-2112878249bd-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7cd4a7e0-6aa4-4418-889a-2112878249bd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.301750 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2jb2\" (UniqueName: \"kubernetes.io/projected/7cd4a7e0-6aa4-4418-889a-2112878249bd-kube-api-access-l2jb2\") pod \"nova-cell1-novncproxy-0\" (UID: \"7cd4a7e0-6aa4-4418-889a-2112878249bd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.307735 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.367218 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.376169 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.573669 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-x8vmm"] Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.618631 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z46bq"] Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.622710 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-z46bq" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.624559 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.624707 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.625873 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z46bq"] Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.661684 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x8vmm" event={"ID":"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958","Type":"ContainerStarted","Data":"f7ff67e412acca047758c87a36fbea8908ac8633261911c16474a3cf22ddcece"} Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.685253 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 16:59:44 crc kubenswrapper[4808]: W1002 16:59:44.687630 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4b16f51_ca0e_41eb_9033_86a5f7ee6091.slice/crio-9ccfe9d28bd411d3138d5f3e9e40ff85060232033069ac6edf9d314714f03ca8 WatchSource:0}: Error finding container 9ccfe9d28bd411d3138d5f3e9e40ff85060232033069ac6edf9d314714f03ca8: Status 404 returned error can't find the container with id 9ccfe9d28bd411d3138d5f3e9e40ff85060232033069ac6edf9d314714f03ca8 Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.721596 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 16:59:44 crc kubenswrapper[4808]: W1002 16:59:44.730132 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7d214c5_98b6_46db_b77a_5415a89227d0.slice/crio-88be6cca977ca1ce0f1ba253ac528b6d89399b793ab2dba204432edc1187be91 WatchSource:0}: Error finding container 88be6cca977ca1ce0f1ba253ac528b6d89399b793ab2dba204432edc1187be91: Status 404 returned error can't find the container with id 88be6cca977ca1ce0f1ba253ac528b6d89399b793ab2dba204432edc1187be91 Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.795160 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-z46bq\" (UID: \"d4c1d967-acf9-4dbb-83db-1f654d4be6d2\") " pod="openstack/nova-cell1-conductor-db-sync-z46bq" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.795240 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-scripts\") pod \"nova-cell1-conductor-db-sync-z46bq\" (UID: \"d4c1d967-acf9-4dbb-83db-1f654d4be6d2\") " pod="openstack/nova-cell1-conductor-db-sync-z46bq" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.795295 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsm5g\" (UniqueName: \"kubernetes.io/projected/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-kube-api-access-dsm5g\") pod \"nova-cell1-conductor-db-sync-z46bq\" (UID: \"d4c1d967-acf9-4dbb-83db-1f654d4be6d2\") " pod="openstack/nova-cell1-conductor-db-sync-z46bq" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.795590 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-config-data\") pod \"nova-cell1-conductor-db-sync-z46bq\" (UID: \"d4c1d967-acf9-4dbb-83db-1f654d4be6d2\") " pod="openstack/nova-cell1-conductor-db-sync-z46bq" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.837633 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.898791 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsm5g\" (UniqueName: \"kubernetes.io/projected/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-kube-api-access-dsm5g\") pod \"nova-cell1-conductor-db-sync-z46bq\" (UID: \"d4c1d967-acf9-4dbb-83db-1f654d4be6d2\") " pod="openstack/nova-cell1-conductor-db-sync-z46bq" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.899545 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-config-data\") pod \"nova-cell1-conductor-db-sync-z46bq\" (UID: \"d4c1d967-acf9-4dbb-83db-1f654d4be6d2\") " pod="openstack/nova-cell1-conductor-db-sync-z46bq" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.899649 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-z46bq\" (UID: \"d4c1d967-acf9-4dbb-83db-1f654d4be6d2\") " pod="openstack/nova-cell1-conductor-db-sync-z46bq" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.899772 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-scripts\") pod \"nova-cell1-conductor-db-sync-z46bq\" (UID: \"d4c1d967-acf9-4dbb-83db-1f654d4be6d2\") " pod="openstack/nova-cell1-conductor-db-sync-z46bq" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.905475 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-scripts\") pod \"nova-cell1-conductor-db-sync-z46bq\" (UID: \"d4c1d967-acf9-4dbb-83db-1f654d4be6d2\") " pod="openstack/nova-cell1-conductor-db-sync-z46bq" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.906017 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-z46bq\" (UID: \"d4c1d967-acf9-4dbb-83db-1f654d4be6d2\") " pod="openstack/nova-cell1-conductor-db-sync-z46bq" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.916423 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsm5g\" (UniqueName: \"kubernetes.io/projected/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-kube-api-access-dsm5g\") pod \"nova-cell1-conductor-db-sync-z46bq\" (UID: \"d4c1d967-acf9-4dbb-83db-1f654d4be6d2\") " pod="openstack/nova-cell1-conductor-db-sync-z46bq" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.917148 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-config-data\") pod \"nova-cell1-conductor-db-sync-z46bq\" (UID: \"d4c1d967-acf9-4dbb-83db-1f654d4be6d2\") " pod="openstack/nova-cell1-conductor-db-sync-z46bq" Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.942808 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-85kbf"] Oct 02 16:59:44 crc kubenswrapper[4808]: I1002 16:59:44.949940 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-z46bq" Oct 02 16:59:45 crc kubenswrapper[4808]: I1002 16:59:45.043895 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 16:59:45 crc kubenswrapper[4808]: I1002 16:59:45.413275 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z46bq"] Oct 02 16:59:45 crc kubenswrapper[4808]: W1002 16:59:45.427092 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4c1d967_acf9_4dbb_83db_1f654d4be6d2.slice/crio-fbdb3f7bec488856e35b21a8b6eb6ac008ee8f3ddd161a7e138e9932c45290d3 WatchSource:0}: Error finding container fbdb3f7bec488856e35b21a8b6eb6ac008ee8f3ddd161a7e138e9932c45290d3: Status 404 returned error can't find the container with id fbdb3f7bec488856e35b21a8b6eb6ac008ee8f3ddd161a7e138e9932c45290d3 Oct 02 16:59:45 crc kubenswrapper[4808]: I1002 16:59:45.674762 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b4b16f51-ca0e-41eb-9033-86a5f7ee6091","Type":"ContainerStarted","Data":"9ccfe9d28bd411d3138d5f3e9e40ff85060232033069ac6edf9d314714f03ca8"} Oct 02 16:59:45 crc kubenswrapper[4808]: I1002 16:59:45.678380 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6c081cab-30c8-4952-8527-65355562b158","Type":"ContainerStarted","Data":"5382cf4df4e928ef18d9ff4685f5fc3678b23786128e30c7d109fa51200d34b4"} Oct 02 16:59:45 crc kubenswrapper[4808]: I1002 16:59:45.679676 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7cd4a7e0-6aa4-4418-889a-2112878249bd","Type":"ContainerStarted","Data":"3cfe8b5418e423e672ce59e3e16d26333e15bbbe203c3ed3a46d693aca1e3225"} Oct 02 16:59:45 crc kubenswrapper[4808]: I1002 16:59:45.681357 4808 generic.go:334] "Generic (PLEG): container finished" podID="fe4d6d85-749d-4ca8-a962-e49678672538" containerID="6939b110451bfa0fdeb8fd14f58f6a84cfc2ce05af2949ca6015d2431e55fe8f" exitCode=0 Oct 02 16:59:45 crc kubenswrapper[4808]: I1002 16:59:45.681399 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-85kbf" event={"ID":"fe4d6d85-749d-4ca8-a962-e49678672538","Type":"ContainerDied","Data":"6939b110451bfa0fdeb8fd14f58f6a84cfc2ce05af2949ca6015d2431e55fe8f"} Oct 02 16:59:45 crc kubenswrapper[4808]: I1002 16:59:45.681415 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-85kbf" event={"ID":"fe4d6d85-749d-4ca8-a962-e49678672538","Type":"ContainerStarted","Data":"f00f248e9e91c2f987a5aec375c0ce40392e5d5d4ac539b2febbc43f32487e0e"} Oct 02 16:59:45 crc kubenswrapper[4808]: I1002 16:59:45.684713 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x8vmm" event={"ID":"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958","Type":"ContainerStarted","Data":"5abae3567d2d28a02690671b1b9a5dd593320c735a0a0c26740bf2573efe3550"} Oct 02 16:59:45 crc kubenswrapper[4808]: I1002 16:59:45.686934 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f7d214c5-98b6-46db-b77a-5415a89227d0","Type":"ContainerStarted","Data":"88be6cca977ca1ce0f1ba253ac528b6d89399b793ab2dba204432edc1187be91"} Oct 02 16:59:45 crc kubenswrapper[4808]: I1002 16:59:45.688637 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-z46bq" event={"ID":"d4c1d967-acf9-4dbb-83db-1f654d4be6d2","Type":"ContainerStarted","Data":"ba2febda635376c978928d84a09d495a559f1728caea409fb8f0bd0d013ea51e"} Oct 02 16:59:45 crc kubenswrapper[4808]: I1002 16:59:45.688662 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-z46bq" event={"ID":"d4c1d967-acf9-4dbb-83db-1f654d4be6d2","Type":"ContainerStarted","Data":"fbdb3f7bec488856e35b21a8b6eb6ac008ee8f3ddd161a7e138e9932c45290d3"} Oct 02 16:59:45 crc kubenswrapper[4808]: I1002 16:59:45.756278 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-x8vmm" podStartSLOduration=2.756258438 podStartE2EDuration="2.756258438s" podCreationTimestamp="2025-10-02 16:59:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:59:45.738719772 +0000 UTC m=+1153.064248792" watchObservedRunningTime="2025-10-02 16:59:45.756258438 +0000 UTC m=+1153.081787438" Oct 02 16:59:45 crc kubenswrapper[4808]: I1002 16:59:45.757097 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-z46bq" podStartSLOduration=1.7570913209999999 podStartE2EDuration="1.757091321s" podCreationTimestamp="2025-10-02 16:59:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:59:45.752162697 +0000 UTC m=+1153.077691697" watchObservedRunningTime="2025-10-02 16:59:45.757091321 +0000 UTC m=+1153.082620321" Oct 02 16:59:47 crc kubenswrapper[4808]: I1002 16:59:47.260676 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 16:59:47 crc kubenswrapper[4808]: I1002 16:59:47.267407 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 16:59:48 crc kubenswrapper[4808]: I1002 16:59:48.714321 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b4b16f51-ca0e-41eb-9033-86a5f7ee6091","Type":"ContainerStarted","Data":"96d3a2743013e5b72aaab1a8e1bf4ba1d55a8160eabc25b4cdf5262c3a5c0fba"} Oct 02 16:59:48 crc kubenswrapper[4808]: I1002 16:59:48.716581 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6c081cab-30c8-4952-8527-65355562b158","Type":"ContainerStarted","Data":"cafa5d95dfc4937e0adc6b08c59a5d0654e48df9151bc9ffb7dc1eeafb0c2dba"} Oct 02 16:59:48 crc kubenswrapper[4808]: I1002 16:59:48.716616 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6c081cab-30c8-4952-8527-65355562b158","Type":"ContainerStarted","Data":"9978bc5c80a90a82c85e7d8655a1c7db8d94b3bd63746bf1c6fefae09a8a29b8"} Oct 02 16:59:48 crc kubenswrapper[4808]: I1002 16:59:48.718023 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7cd4a7e0-6aa4-4418-889a-2112878249bd","Type":"ContainerStarted","Data":"89aeb4bdb27b5d5cc32c7d7802c649112b39faa31e0ffae6369762cd3ac6f778"} Oct 02 16:59:48 crc kubenswrapper[4808]: I1002 16:59:48.718125 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="7cd4a7e0-6aa4-4418-889a-2112878249bd" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://89aeb4bdb27b5d5cc32c7d7802c649112b39faa31e0ffae6369762cd3ac6f778" gracePeriod=30 Oct 02 16:59:48 crc kubenswrapper[4808]: I1002 16:59:48.720753 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-85kbf" event={"ID":"fe4d6d85-749d-4ca8-a962-e49678672538","Type":"ContainerStarted","Data":"14265a8079d6478a47893642811ce11637483a21a6d50eb089c39f67ea514e48"} Oct 02 16:59:48 crc kubenswrapper[4808]: I1002 16:59:48.720914 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 16:59:48 crc kubenswrapper[4808]: I1002 16:59:48.726597 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f7d214c5-98b6-46db-b77a-5415a89227d0","Type":"ContainerStarted","Data":"e5ca8426cead7b8e3e7cfc5c34bcf880e3203eb682cce840135810b487909c93"} Oct 02 16:59:48 crc kubenswrapper[4808]: I1002 16:59:48.726638 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f7d214c5-98b6-46db-b77a-5415a89227d0","Type":"ContainerStarted","Data":"5a78fd379316675deb050b865bda7acda96094ea1081fcd33dd1bee0b8318064"} Oct 02 16:59:48 crc kubenswrapper[4808]: I1002 16:59:48.726703 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f7d214c5-98b6-46db-b77a-5415a89227d0" containerName="nova-metadata-log" containerID="cri-o://5a78fd379316675deb050b865bda7acda96094ea1081fcd33dd1bee0b8318064" gracePeriod=30 Oct 02 16:59:48 crc kubenswrapper[4808]: I1002 16:59:48.726743 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f7d214c5-98b6-46db-b77a-5415a89227d0" containerName="nova-metadata-metadata" containerID="cri-o://e5ca8426cead7b8e3e7cfc5c34bcf880e3203eb682cce840135810b487909c93" gracePeriod=30 Oct 02 16:59:48 crc kubenswrapper[4808]: I1002 16:59:48.743448 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.595481874 podStartE2EDuration="5.74342901s" podCreationTimestamp="2025-10-02 16:59:43 +0000 UTC" firstStartedPulling="2025-10-02 16:59:44.690137306 +0000 UTC m=+1152.015666306" lastFinishedPulling="2025-10-02 16:59:47.838084402 +0000 UTC m=+1155.163613442" observedRunningTime="2025-10-02 16:59:48.728301449 +0000 UTC m=+1156.053830449" watchObservedRunningTime="2025-10-02 16:59:48.74342901 +0000 UTC m=+1156.068958010" Oct 02 16:59:48 crc kubenswrapper[4808]: I1002 16:59:48.750449 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.5714368309999998 podStartE2EDuration="5.75041766s" podCreationTimestamp="2025-10-02 16:59:43 +0000 UTC" firstStartedPulling="2025-10-02 16:59:44.732092995 +0000 UTC m=+1152.057621995" lastFinishedPulling="2025-10-02 16:59:47.911073784 +0000 UTC m=+1155.236602824" observedRunningTime="2025-10-02 16:59:48.749630648 +0000 UTC m=+1156.075159648" watchObservedRunningTime="2025-10-02 16:59:48.75041766 +0000 UTC m=+1156.075946660" Oct 02 16:59:48 crc kubenswrapper[4808]: I1002 16:59:48.771732 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.906831706 podStartE2EDuration="5.771714498s" podCreationTimestamp="2025-10-02 16:59:43 +0000 UTC" firstStartedPulling="2025-10-02 16:59:45.044462035 +0000 UTC m=+1152.369991035" lastFinishedPulling="2025-10-02 16:59:47.909344787 +0000 UTC m=+1155.234873827" observedRunningTime="2025-10-02 16:59:48.769106367 +0000 UTC m=+1156.094635367" watchObservedRunningTime="2025-10-02 16:59:48.771714498 +0000 UTC m=+1156.097243498" Oct 02 16:59:48 crc kubenswrapper[4808]: I1002 16:59:48.804976 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.773663 podStartE2EDuration="5.80496138s" podCreationTimestamp="2025-10-02 16:59:43 +0000 UTC" firstStartedPulling="2025-10-02 16:59:44.838944895 +0000 UTC m=+1152.164473895" lastFinishedPulling="2025-10-02 16:59:47.870243265 +0000 UTC m=+1155.195772275" observedRunningTime="2025-10-02 16:59:48.799011259 +0000 UTC m=+1156.124540259" watchObservedRunningTime="2025-10-02 16:59:48.80496138 +0000 UTC m=+1156.130490380" Oct 02 16:59:48 crc kubenswrapper[4808]: I1002 16:59:48.827118 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-566b5b7845-85kbf" podStartSLOduration=5.827097981 podStartE2EDuration="5.827097981s" podCreationTimestamp="2025-10-02 16:59:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:59:48.817138931 +0000 UTC m=+1156.142667931" watchObservedRunningTime="2025-10-02 16:59:48.827097981 +0000 UTC m=+1156.152626981" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.220324 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.220376 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.298605 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.308106 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.377308 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.413860 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7d214c5-98b6-46db-b77a-5415a89227d0-combined-ca-bundle\") pod \"f7d214c5-98b6-46db-b77a-5415a89227d0\" (UID: \"f7d214c5-98b6-46db-b77a-5415a89227d0\") " Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.413954 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndgp2\" (UniqueName: \"kubernetes.io/projected/f7d214c5-98b6-46db-b77a-5415a89227d0-kube-api-access-ndgp2\") pod \"f7d214c5-98b6-46db-b77a-5415a89227d0\" (UID: \"f7d214c5-98b6-46db-b77a-5415a89227d0\") " Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.413998 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7d214c5-98b6-46db-b77a-5415a89227d0-config-data\") pod \"f7d214c5-98b6-46db-b77a-5415a89227d0\" (UID: \"f7d214c5-98b6-46db-b77a-5415a89227d0\") " Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.414024 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7d214c5-98b6-46db-b77a-5415a89227d0-logs\") pod \"f7d214c5-98b6-46db-b77a-5415a89227d0\" (UID: \"f7d214c5-98b6-46db-b77a-5415a89227d0\") " Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.414707 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7d214c5-98b6-46db-b77a-5415a89227d0-logs" (OuterVolumeSpecName: "logs") pod "f7d214c5-98b6-46db-b77a-5415a89227d0" (UID: "f7d214c5-98b6-46db-b77a-5415a89227d0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.428958 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7d214c5-98b6-46db-b77a-5415a89227d0-kube-api-access-ndgp2" (OuterVolumeSpecName: "kube-api-access-ndgp2") pod "f7d214c5-98b6-46db-b77a-5415a89227d0" (UID: "f7d214c5-98b6-46db-b77a-5415a89227d0"). InnerVolumeSpecName "kube-api-access-ndgp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.459068 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7d214c5-98b6-46db-b77a-5415a89227d0-config-data" (OuterVolumeSpecName: "config-data") pod "f7d214c5-98b6-46db-b77a-5415a89227d0" (UID: "f7d214c5-98b6-46db-b77a-5415a89227d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.463105 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7d214c5-98b6-46db-b77a-5415a89227d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7d214c5-98b6-46db-b77a-5415a89227d0" (UID: "f7d214c5-98b6-46db-b77a-5415a89227d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.515655 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7d214c5-98b6-46db-b77a-5415a89227d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.515685 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndgp2\" (UniqueName: \"kubernetes.io/projected/f7d214c5-98b6-46db-b77a-5415a89227d0-kube-api-access-ndgp2\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.515697 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7d214c5-98b6-46db-b77a-5415a89227d0-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.515705 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7d214c5-98b6-46db-b77a-5415a89227d0-logs\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.737489 4808 generic.go:334] "Generic (PLEG): container finished" podID="f7d214c5-98b6-46db-b77a-5415a89227d0" containerID="e5ca8426cead7b8e3e7cfc5c34bcf880e3203eb682cce840135810b487909c93" exitCode=0 Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.737526 4808 generic.go:334] "Generic (PLEG): container finished" podID="f7d214c5-98b6-46db-b77a-5415a89227d0" containerID="5a78fd379316675deb050b865bda7acda96094ea1081fcd33dd1bee0b8318064" exitCode=143 Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.737547 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.737616 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f7d214c5-98b6-46db-b77a-5415a89227d0","Type":"ContainerDied","Data":"e5ca8426cead7b8e3e7cfc5c34bcf880e3203eb682cce840135810b487909c93"} Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.737646 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f7d214c5-98b6-46db-b77a-5415a89227d0","Type":"ContainerDied","Data":"5a78fd379316675deb050b865bda7acda96094ea1081fcd33dd1bee0b8318064"} Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.737660 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f7d214c5-98b6-46db-b77a-5415a89227d0","Type":"ContainerDied","Data":"88be6cca977ca1ce0f1ba253ac528b6d89399b793ab2dba204432edc1187be91"} Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.737678 4808 scope.go:117] "RemoveContainer" containerID="e5ca8426cead7b8e3e7cfc5c34bcf880e3203eb682cce840135810b487909c93" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.783463 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.795323 4808 scope.go:117] "RemoveContainer" containerID="5a78fd379316675deb050b865bda7acda96094ea1081fcd33dd1bee0b8318064" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.809919 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.818136 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 16:59:49 crc kubenswrapper[4808]: E1002 16:59:49.818815 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7d214c5-98b6-46db-b77a-5415a89227d0" containerName="nova-metadata-metadata" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.818880 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7d214c5-98b6-46db-b77a-5415a89227d0" containerName="nova-metadata-metadata" Oct 02 16:59:49 crc kubenswrapper[4808]: E1002 16:59:49.818936 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7d214c5-98b6-46db-b77a-5415a89227d0" containerName="nova-metadata-log" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.818950 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7d214c5-98b6-46db-b77a-5415a89227d0" containerName="nova-metadata-log" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.819317 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7d214c5-98b6-46db-b77a-5415a89227d0" containerName="nova-metadata-log" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.819373 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7d214c5-98b6-46db-b77a-5415a89227d0" containerName="nova-metadata-metadata" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.820954 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.825583 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.827627 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.838370 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.851253 4808 scope.go:117] "RemoveContainer" containerID="e5ca8426cead7b8e3e7cfc5c34bcf880e3203eb682cce840135810b487909c93" Oct 02 16:59:49 crc kubenswrapper[4808]: E1002 16:59:49.851811 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5ca8426cead7b8e3e7cfc5c34bcf880e3203eb682cce840135810b487909c93\": container with ID starting with e5ca8426cead7b8e3e7cfc5c34bcf880e3203eb682cce840135810b487909c93 not found: ID does not exist" containerID="e5ca8426cead7b8e3e7cfc5c34bcf880e3203eb682cce840135810b487909c93" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.851850 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5ca8426cead7b8e3e7cfc5c34bcf880e3203eb682cce840135810b487909c93"} err="failed to get container status \"e5ca8426cead7b8e3e7cfc5c34bcf880e3203eb682cce840135810b487909c93\": rpc error: code = NotFound desc = could not find container \"e5ca8426cead7b8e3e7cfc5c34bcf880e3203eb682cce840135810b487909c93\": container with ID starting with e5ca8426cead7b8e3e7cfc5c34bcf880e3203eb682cce840135810b487909c93 not found: ID does not exist" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.851870 4808 scope.go:117] "RemoveContainer" containerID="5a78fd379316675deb050b865bda7acda96094ea1081fcd33dd1bee0b8318064" Oct 02 16:59:49 crc kubenswrapper[4808]: E1002 16:59:49.852158 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a78fd379316675deb050b865bda7acda96094ea1081fcd33dd1bee0b8318064\": container with ID starting with 5a78fd379316675deb050b865bda7acda96094ea1081fcd33dd1bee0b8318064 not found: ID does not exist" containerID="5a78fd379316675deb050b865bda7acda96094ea1081fcd33dd1bee0b8318064" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.852182 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a78fd379316675deb050b865bda7acda96094ea1081fcd33dd1bee0b8318064"} err="failed to get container status \"5a78fd379316675deb050b865bda7acda96094ea1081fcd33dd1bee0b8318064\": rpc error: code = NotFound desc = could not find container \"5a78fd379316675deb050b865bda7acda96094ea1081fcd33dd1bee0b8318064\": container with ID starting with 5a78fd379316675deb050b865bda7acda96094ea1081fcd33dd1bee0b8318064 not found: ID does not exist" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.852198 4808 scope.go:117] "RemoveContainer" containerID="e5ca8426cead7b8e3e7cfc5c34bcf880e3203eb682cce840135810b487909c93" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.852604 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5ca8426cead7b8e3e7cfc5c34bcf880e3203eb682cce840135810b487909c93"} err="failed to get container status \"e5ca8426cead7b8e3e7cfc5c34bcf880e3203eb682cce840135810b487909c93\": rpc error: code = NotFound desc = could not find container \"e5ca8426cead7b8e3e7cfc5c34bcf880e3203eb682cce840135810b487909c93\": container with ID starting with e5ca8426cead7b8e3e7cfc5c34bcf880e3203eb682cce840135810b487909c93 not found: ID does not exist" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.852628 4808 scope.go:117] "RemoveContainer" containerID="5a78fd379316675deb050b865bda7acda96094ea1081fcd33dd1bee0b8318064" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.852884 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a78fd379316675deb050b865bda7acda96094ea1081fcd33dd1bee0b8318064"} err="failed to get container status \"5a78fd379316675deb050b865bda7acda96094ea1081fcd33dd1bee0b8318064\": rpc error: code = NotFound desc = could not find container \"5a78fd379316675deb050b865bda7acda96094ea1081fcd33dd1bee0b8318064\": container with ID starting with 5a78fd379316675deb050b865bda7acda96094ea1081fcd33dd1bee0b8318064 not found: ID does not exist" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.924406 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\") " pod="openstack/nova-metadata-0" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.924475 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4lpc\" (UniqueName: \"kubernetes.io/projected/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-kube-api-access-t4lpc\") pod \"nova-metadata-0\" (UID: \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\") " pod="openstack/nova-metadata-0" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.924504 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-config-data\") pod \"nova-metadata-0\" (UID: \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\") " pod="openstack/nova-metadata-0" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.924553 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-logs\") pod \"nova-metadata-0\" (UID: \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\") " pod="openstack/nova-metadata-0" Oct 02 16:59:49 crc kubenswrapper[4808]: I1002 16:59:49.924641 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\") " pod="openstack/nova-metadata-0" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.025985 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\") " pod="openstack/nova-metadata-0" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.026087 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\") " pod="openstack/nova-metadata-0" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.026138 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4lpc\" (UniqueName: \"kubernetes.io/projected/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-kube-api-access-t4lpc\") pod \"nova-metadata-0\" (UID: \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\") " pod="openstack/nova-metadata-0" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.026166 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-config-data\") pod \"nova-metadata-0\" (UID: \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\") " pod="openstack/nova-metadata-0" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.026220 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-logs\") pod \"nova-metadata-0\" (UID: \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\") " pod="openstack/nova-metadata-0" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.026657 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-logs\") pod \"nova-metadata-0\" (UID: \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\") " pod="openstack/nova-metadata-0" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.031407 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\") " pod="openstack/nova-metadata-0" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.031765 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-config-data\") pod \"nova-metadata-0\" (UID: \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\") " pod="openstack/nova-metadata-0" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.037576 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\") " pod="openstack/nova-metadata-0" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.052312 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4lpc\" (UniqueName: \"kubernetes.io/projected/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-kube-api-access-t4lpc\") pod \"nova-metadata-0\" (UID: \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\") " pod="openstack/nova-metadata-0" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.203809 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.204367 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="f18f563e-79e6-4c33-87a5-91f7c27bf181" containerName="kube-state-metrics" containerID="cri-o://b6e973576609e6ccbd8861e21134aebcda316cfd3aa63914174cc4babd323e49" gracePeriod=30 Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.211396 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.472402 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="f18f563e-79e6-4c33-87a5-91f7c27bf181" containerName="kube-state-metrics" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": dial tcp 10.217.0.105:8081: connect: connection refused" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.662025 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.736393 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.737008 4808 scope.go:117] "RemoveContainer" containerID="a918ee069243e5dc0b7673d53697e23876162204008e5ae0c51065fa9012a203" Oct 02 16:59:50 crc kubenswrapper[4808]: E1002 16:59:50.737218 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=keystone-operator-controller-manager-65f4bcbb94-svkdd_openstack-operators(818e02a3-653a-4b36-a68b-9d4e2df0b71e)\"" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" podUID="818e02a3-653a-4b36-a68b-9d4e2df0b71e" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.763777 4808 generic.go:334] "Generic (PLEG): container finished" podID="f18f563e-79e6-4c33-87a5-91f7c27bf181" containerID="b6e973576609e6ccbd8861e21134aebcda316cfd3aa63914174cc4babd323e49" exitCode=2 Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.763884 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f18f563e-79e6-4c33-87a5-91f7c27bf181","Type":"ContainerDied","Data":"b6e973576609e6ccbd8861e21134aebcda316cfd3aa63914174cc4babd323e49"} Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.763914 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f18f563e-79e6-4c33-87a5-91f7c27bf181","Type":"ContainerDied","Data":"d02e37733c60add96e9eafa89bb29e873dfd6102c813fe0f86fbaed5ec4cfb96"} Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.763935 4808 scope.go:117] "RemoveContainer" containerID="b6e973576609e6ccbd8861e21134aebcda316cfd3aa63914174cc4babd323e49" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.764000 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.802392 4808 scope.go:117] "RemoveContainer" containerID="b6e973576609e6ccbd8861e21134aebcda316cfd3aa63914174cc4babd323e49" Oct 02 16:59:50 crc kubenswrapper[4808]: E1002 16:59:50.802958 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6e973576609e6ccbd8861e21134aebcda316cfd3aa63914174cc4babd323e49\": container with ID starting with b6e973576609e6ccbd8861e21134aebcda316cfd3aa63914174cc4babd323e49 not found: ID does not exist" containerID="b6e973576609e6ccbd8861e21134aebcda316cfd3aa63914174cc4babd323e49" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.803008 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6e973576609e6ccbd8861e21134aebcda316cfd3aa63914174cc4babd323e49"} err="failed to get container status \"b6e973576609e6ccbd8861e21134aebcda316cfd3aa63914174cc4babd323e49\": rpc error: code = NotFound desc = could not find container \"b6e973576609e6ccbd8861e21134aebcda316cfd3aa63914174cc4babd323e49\": container with ID starting with b6e973576609e6ccbd8861e21134aebcda316cfd3aa63914174cc4babd323e49 not found: ID does not exist" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.803669 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.847110 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lr2wm\" (UniqueName: \"kubernetes.io/projected/f18f563e-79e6-4c33-87a5-91f7c27bf181-kube-api-access-lr2wm\") pod \"f18f563e-79e6-4c33-87a5-91f7c27bf181\" (UID: \"f18f563e-79e6-4c33-87a5-91f7c27bf181\") " Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.852497 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f18f563e-79e6-4c33-87a5-91f7c27bf181-kube-api-access-lr2wm" (OuterVolumeSpecName: "kube-api-access-lr2wm") pod "f18f563e-79e6-4c33-87a5-91f7c27bf181" (UID: "f18f563e-79e6-4c33-87a5-91f7c27bf181"). InnerVolumeSpecName "kube-api-access-lr2wm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:59:50 crc kubenswrapper[4808]: I1002 16:59:50.949011 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lr2wm\" (UniqueName: \"kubernetes.io/projected/f18f563e-79e6-4c33-87a5-91f7c27bf181-kube-api-access-lr2wm\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.215979 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.228481 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.249563 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 16:59:51 crc kubenswrapper[4808]: E1002 16:59:51.250228 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f18f563e-79e6-4c33-87a5-91f7c27bf181" containerName="kube-state-metrics" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.250253 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f18f563e-79e6-4c33-87a5-91f7c27bf181" containerName="kube-state-metrics" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.250643 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f18f563e-79e6-4c33-87a5-91f7c27bf181" containerName="kube-state-metrics" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.251781 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.263171 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.263221 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.277009 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.344357 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.345100 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" containerName="ceilometer-central-agent" containerID="cri-o://80ed9069762936ea68104c7c75c37bfd9b082fa028cd3d54e34778db6bf53a6a" gracePeriod=30 Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.345667 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" containerName="proxy-httpd" containerID="cri-o://fe7ced19e6b556a1b7a881eaf296ab31a77bc3a70e065fde7fd3dd34a1714fe9" gracePeriod=30 Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.345779 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" containerName="sg-core" containerID="cri-o://94522ef9e646b66e17dc4bfa2f54fe131a24052e954ad1d679f5b93845992bf0" gracePeriod=30 Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.345705 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" containerName="ceilometer-notification-agent" containerID="cri-o://3d0dd557125f47597903502f7aea47322f8712e52bf3c8140a68ed8715fe94da" gracePeriod=30 Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.361402 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/93c29105-8d7b-4104-9d99-b5e4bd697253-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"93c29105-8d7b-4104-9d99-b5e4bd697253\") " pod="openstack/kube-state-metrics-0" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.361467 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/93c29105-8d7b-4104-9d99-b5e4bd697253-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"93c29105-8d7b-4104-9d99-b5e4bd697253\") " pod="openstack/kube-state-metrics-0" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.361492 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93c29105-8d7b-4104-9d99-b5e4bd697253-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"93c29105-8d7b-4104-9d99-b5e4bd697253\") " pod="openstack/kube-state-metrics-0" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.361531 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g562\" (UniqueName: \"kubernetes.io/projected/93c29105-8d7b-4104-9d99-b5e4bd697253-kube-api-access-6g562\") pod \"kube-state-metrics-0\" (UID: \"93c29105-8d7b-4104-9d99-b5e4bd697253\") " pod="openstack/kube-state-metrics-0" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.406074 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f18f563e-79e6-4c33-87a5-91f7c27bf181" path="/var/lib/kubelet/pods/f18f563e-79e6-4c33-87a5-91f7c27bf181/volumes" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.406614 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7d214c5-98b6-46db-b77a-5415a89227d0" path="/var/lib/kubelet/pods/f7d214c5-98b6-46db-b77a-5415a89227d0/volumes" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.462835 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/93c29105-8d7b-4104-9d99-b5e4bd697253-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"93c29105-8d7b-4104-9d99-b5e4bd697253\") " pod="openstack/kube-state-metrics-0" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.462907 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/93c29105-8d7b-4104-9d99-b5e4bd697253-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"93c29105-8d7b-4104-9d99-b5e4bd697253\") " pod="openstack/kube-state-metrics-0" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.462944 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93c29105-8d7b-4104-9d99-b5e4bd697253-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"93c29105-8d7b-4104-9d99-b5e4bd697253\") " pod="openstack/kube-state-metrics-0" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.462974 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g562\" (UniqueName: \"kubernetes.io/projected/93c29105-8d7b-4104-9d99-b5e4bd697253-kube-api-access-6g562\") pod \"kube-state-metrics-0\" (UID: \"93c29105-8d7b-4104-9d99-b5e4bd697253\") " pod="openstack/kube-state-metrics-0" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.465756 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/93c29105-8d7b-4104-9d99-b5e4bd697253-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"93c29105-8d7b-4104-9d99-b5e4bd697253\") " pod="openstack/kube-state-metrics-0" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.466082 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93c29105-8d7b-4104-9d99-b5e4bd697253-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"93c29105-8d7b-4104-9d99-b5e4bd697253\") " pod="openstack/kube-state-metrics-0" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.466371 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/93c29105-8d7b-4104-9d99-b5e4bd697253-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"93c29105-8d7b-4104-9d99-b5e4bd697253\") " pod="openstack/kube-state-metrics-0" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.479535 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g562\" (UniqueName: \"kubernetes.io/projected/93c29105-8d7b-4104-9d99-b5e4bd697253-kube-api-access-6g562\") pod \"kube-state-metrics-0\" (UID: \"93c29105-8d7b-4104-9d99-b5e4bd697253\") " pod="openstack/kube-state-metrics-0" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.581016 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.788261 4808 generic.go:334] "Generic (PLEG): container finished" podID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" containerID="fe7ced19e6b556a1b7a881eaf296ab31a77bc3a70e065fde7fd3dd34a1714fe9" exitCode=0 Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.788535 4808 generic.go:334] "Generic (PLEG): container finished" podID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" containerID="94522ef9e646b66e17dc4bfa2f54fe131a24052e954ad1d679f5b93845992bf0" exitCode=2 Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.788546 4808 generic.go:334] "Generic (PLEG): container finished" podID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" containerID="80ed9069762936ea68104c7c75c37bfd9b082fa028cd3d54e34778db6bf53a6a" exitCode=0 Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.788588 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7","Type":"ContainerDied","Data":"fe7ced19e6b556a1b7a881eaf296ab31a77bc3a70e065fde7fd3dd34a1714fe9"} Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.788617 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7","Type":"ContainerDied","Data":"94522ef9e646b66e17dc4bfa2f54fe131a24052e954ad1d679f5b93845992bf0"} Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.788629 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7","Type":"ContainerDied","Data":"80ed9069762936ea68104c7c75c37bfd9b082fa028cd3d54e34778db6bf53a6a"} Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.790481 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e","Type":"ContainerStarted","Data":"87868c82af98dc65f5e0b08a41f6da75a7048153f3aa5c476f11215d1c792da6"} Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.790511 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e","Type":"ContainerStarted","Data":"fac71ac72b0bd975678a5f8f609d60cba7908c936f1ef0df73f8d09495411ec0"} Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.790546 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e","Type":"ContainerStarted","Data":"8014d5ac162322b414aefcc3257685427d75432fe10755b72325b8d8ae1abb8c"} Oct 02 16:59:51 crc kubenswrapper[4808]: I1002 16:59:51.827490 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.827468712 podStartE2EDuration="2.827468712s" podCreationTimestamp="2025-10-02 16:59:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:59:51.823927906 +0000 UTC m=+1159.149456926" watchObservedRunningTime="2025-10-02 16:59:51.827468712 +0000 UTC m=+1159.152997712" Oct 02 16:59:52 crc kubenswrapper[4808]: I1002 16:59:52.004998 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 16:59:52 crc kubenswrapper[4808]: I1002 16:59:52.804226 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"93c29105-8d7b-4104-9d99-b5e4bd697253","Type":"ContainerStarted","Data":"0494e6c11d979df86df599e87e58656b14b9f869dfe1a7a237adf684745d194b"} Oct 02 16:59:52 crc kubenswrapper[4808]: I1002 16:59:52.804598 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"93c29105-8d7b-4104-9d99-b5e4bd697253","Type":"ContainerStarted","Data":"b8482b1005a2112ba31116769518886795e78868118dd7264a4bd53d112d0ff4"} Oct 02 16:59:52 crc kubenswrapper[4808]: I1002 16:59:52.804663 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 16:59:52 crc kubenswrapper[4808]: I1002 16:59:52.806295 4808 generic.go:334] "Generic (PLEG): container finished" podID="c4bb01f7-7c71-4b46-b45d-49d3cd6b0958" containerID="5abae3567d2d28a02690671b1b9a5dd593320c735a0a0c26740bf2573efe3550" exitCode=0 Oct 02 16:59:52 crc kubenswrapper[4808]: I1002 16:59:52.806325 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x8vmm" event={"ID":"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958","Type":"ContainerDied","Data":"5abae3567d2d28a02690671b1b9a5dd593320c735a0a0c26740bf2573efe3550"} Oct 02 16:59:52 crc kubenswrapper[4808]: I1002 16:59:52.831039 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.388127312 podStartE2EDuration="1.831020285s" podCreationTimestamp="2025-10-02 16:59:51 +0000 UTC" firstStartedPulling="2025-10-02 16:59:52.014637413 +0000 UTC m=+1159.340166413" lastFinishedPulling="2025-10-02 16:59:52.457530376 +0000 UTC m=+1159.783059386" observedRunningTime="2025-10-02 16:59:52.822977687 +0000 UTC m=+1160.148506687" watchObservedRunningTime="2025-10-02 16:59:52.831020285 +0000 UTC m=+1160.156549285" Oct 02 16:59:53 crc kubenswrapper[4808]: I1002 16:59:53.819760 4808 generic.go:334] "Generic (PLEG): container finished" podID="d4c1d967-acf9-4dbb-83db-1f654d4be6d2" containerID="ba2febda635376c978928d84a09d495a559f1728caea409fb8f0bd0d013ea51e" exitCode=0 Oct 02 16:59:53 crc kubenswrapper[4808]: I1002 16:59:53.819905 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-z46bq" event={"ID":"d4c1d967-acf9-4dbb-83db-1f654d4be6d2","Type":"ContainerDied","Data":"ba2febda635376c978928d84a09d495a559f1728caea409fb8f0bd0d013ea51e"} Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.182666 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x8vmm" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.268623 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.268665 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.308763 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.318167 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-config-data\") pod \"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958\" (UID: \"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958\") " Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.318435 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-scripts\") pod \"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958\" (UID: \"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958\") " Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.318501 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-combined-ca-bundle\") pod \"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958\" (UID: \"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958\") " Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.318540 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8hfg\" (UniqueName: \"kubernetes.io/projected/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-kube-api-access-r8hfg\") pod \"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958\" (UID: \"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958\") " Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.332425 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-scripts" (OuterVolumeSpecName: "scripts") pod "c4bb01f7-7c71-4b46-b45d-49d3cd6b0958" (UID: "c4bb01f7-7c71-4b46-b45d-49d3cd6b0958"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.332639 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-kube-api-access-r8hfg" (OuterVolumeSpecName: "kube-api-access-r8hfg") pod "c4bb01f7-7c71-4b46-b45d-49d3cd6b0958" (UID: "c4bb01f7-7c71-4b46-b45d-49d3cd6b0958"). InnerVolumeSpecName "kube-api-access-r8hfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.350135 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.369375 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.371856 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-config-data" (OuterVolumeSpecName: "config-data") pod "c4bb01f7-7c71-4b46-b45d-49d3cd6b0958" (UID: "c4bb01f7-7c71-4b46-b45d-49d3cd6b0958"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.376502 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c4bb01f7-7c71-4b46-b45d-49d3cd6b0958" (UID: "c4bb01f7-7c71-4b46-b45d-49d3cd6b0958"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.422007 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.422522 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.422842 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8hfg\" (UniqueName: \"kubernetes.io/projected/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-kube-api-access-r8hfg\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.422981 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.461473 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-g2j2k"] Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.462062 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" podUID="c166b259-1f16-4e18-8df1-a4a222d49897" containerName="dnsmasq-dns" containerID="cri-o://39095e68ee42cf7a1370c4cbb6399805c01b1864ac520be1c9a337fb1d065ba2" gracePeriod=10 Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.824771 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.829119 4808 generic.go:334] "Generic (PLEG): container finished" podID="c166b259-1f16-4e18-8df1-a4a222d49897" containerID="39095e68ee42cf7a1370c4cbb6399805c01b1864ac520be1c9a337fb1d065ba2" exitCode=0 Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.829192 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" event={"ID":"c166b259-1f16-4e18-8df1-a4a222d49897","Type":"ContainerDied","Data":"39095e68ee42cf7a1370c4cbb6399805c01b1864ac520be1c9a337fb1d065ba2"} Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.829222 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" event={"ID":"c166b259-1f16-4e18-8df1-a4a222d49897","Type":"ContainerDied","Data":"c58fbada1c5d03748625c8809ce46ab79b82807a863c677ec91824365e96ba5c"} Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.829260 4808 scope.go:117] "RemoveContainer" containerID="39095e68ee42cf7a1370c4cbb6399805c01b1864ac520be1c9a337fb1d065ba2" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.829406 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-g2j2k" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.832842 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x8vmm" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.833574 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x8vmm" event={"ID":"c4bb01f7-7c71-4b46-b45d-49d3cd6b0958","Type":"ContainerDied","Data":"f7ff67e412acca047758c87a36fbea8908ac8633261911c16474a3cf22ddcece"} Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.833614 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7ff67e412acca047758c87a36fbea8908ac8633261911c16474a3cf22ddcece" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.865209 4808 scope.go:117] "RemoveContainer" containerID="728155e665696413326ab58d71546bb2627e8d71d48745322b77e76a0064a596" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.885601 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.909377 4808 scope.go:117] "RemoveContainer" containerID="39095e68ee42cf7a1370c4cbb6399805c01b1864ac520be1c9a337fb1d065ba2" Oct 02 16:59:54 crc kubenswrapper[4808]: E1002 16:59:54.910832 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39095e68ee42cf7a1370c4cbb6399805c01b1864ac520be1c9a337fb1d065ba2\": container with ID starting with 39095e68ee42cf7a1370c4cbb6399805c01b1864ac520be1c9a337fb1d065ba2 not found: ID does not exist" containerID="39095e68ee42cf7a1370c4cbb6399805c01b1864ac520be1c9a337fb1d065ba2" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.910881 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39095e68ee42cf7a1370c4cbb6399805c01b1864ac520be1c9a337fb1d065ba2"} err="failed to get container status \"39095e68ee42cf7a1370c4cbb6399805c01b1864ac520be1c9a337fb1d065ba2\": rpc error: code = NotFound desc = could not find container \"39095e68ee42cf7a1370c4cbb6399805c01b1864ac520be1c9a337fb1d065ba2\": container with ID starting with 39095e68ee42cf7a1370c4cbb6399805c01b1864ac520be1c9a337fb1d065ba2 not found: ID does not exist" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.910917 4808 scope.go:117] "RemoveContainer" containerID="728155e665696413326ab58d71546bb2627e8d71d48745322b77e76a0064a596" Oct 02 16:59:54 crc kubenswrapper[4808]: E1002 16:59:54.911553 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"728155e665696413326ab58d71546bb2627e8d71d48745322b77e76a0064a596\": container with ID starting with 728155e665696413326ab58d71546bb2627e8d71d48745322b77e76a0064a596 not found: ID does not exist" containerID="728155e665696413326ab58d71546bb2627e8d71d48745322b77e76a0064a596" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.911595 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"728155e665696413326ab58d71546bb2627e8d71d48745322b77e76a0064a596"} err="failed to get container status \"728155e665696413326ab58d71546bb2627e8d71d48745322b77e76a0064a596\": rpc error: code = NotFound desc = could not find container \"728155e665696413326ab58d71546bb2627e8d71d48745322b77e76a0064a596\": container with ID starting with 728155e665696413326ab58d71546bb2627e8d71d48745322b77e76a0064a596 not found: ID does not exist" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.936034 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-ovsdbserver-nb\") pod \"c166b259-1f16-4e18-8df1-a4a222d49897\" (UID: \"c166b259-1f16-4e18-8df1-a4a222d49897\") " Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.936127 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-ovsdbserver-sb\") pod \"c166b259-1f16-4e18-8df1-a4a222d49897\" (UID: \"c166b259-1f16-4e18-8df1-a4a222d49897\") " Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.936202 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-config\") pod \"c166b259-1f16-4e18-8df1-a4a222d49897\" (UID: \"c166b259-1f16-4e18-8df1-a4a222d49897\") " Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.936283 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92ch8\" (UniqueName: \"kubernetes.io/projected/c166b259-1f16-4e18-8df1-a4a222d49897-kube-api-access-92ch8\") pod \"c166b259-1f16-4e18-8df1-a4a222d49897\" (UID: \"c166b259-1f16-4e18-8df1-a4a222d49897\") " Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.936451 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-dns-svc\") pod \"c166b259-1f16-4e18-8df1-a4a222d49897\" (UID: \"c166b259-1f16-4e18-8df1-a4a222d49897\") " Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.942092 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c166b259-1f16-4e18-8df1-a4a222d49897-kube-api-access-92ch8" (OuterVolumeSpecName: "kube-api-access-92ch8") pod "c166b259-1f16-4e18-8df1-a4a222d49897" (UID: "c166b259-1f16-4e18-8df1-a4a222d49897"). InnerVolumeSpecName "kube-api-access-92ch8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.983102 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c166b259-1f16-4e18-8df1-a4a222d49897" (UID: "c166b259-1f16-4e18-8df1-a4a222d49897"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.986738 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-config" (OuterVolumeSpecName: "config") pod "c166b259-1f16-4e18-8df1-a4a222d49897" (UID: "c166b259-1f16-4e18-8df1-a4a222d49897"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:59:54 crc kubenswrapper[4808]: I1002 16:59:54.991836 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c166b259-1f16-4e18-8df1-a4a222d49897" (UID: "c166b259-1f16-4e18-8df1-a4a222d49897"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.012457 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c166b259-1f16-4e18-8df1-a4a222d49897" (UID: "c166b259-1f16-4e18-8df1-a4a222d49897"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.046966 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.047023 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.047034 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-config\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.047321 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92ch8\" (UniqueName: \"kubernetes.io/projected/c166b259-1f16-4e18-8df1-a4a222d49897-kube-api-access-92ch8\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.047337 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c166b259-1f16-4e18-8df1-a4a222d49897-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.067840 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.068086 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6c081cab-30c8-4952-8527-65355562b158" containerName="nova-api-log" containerID="cri-o://9978bc5c80a90a82c85e7d8655a1c7db8d94b3bd63746bf1c6fefae09a8a29b8" gracePeriod=30 Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.068228 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6c081cab-30c8-4952-8527-65355562b158" containerName="nova-api-api" containerID="cri-o://cafa5d95dfc4937e0adc6b08c59a5d0654e48df9151bc9ffb7dc1eeafb0c2dba" gracePeriod=30 Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.082515 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6c081cab-30c8-4952-8527-65355562b158" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.171:8774/\": EOF" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.082514 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6c081cab-30c8-4952-8527-65355562b158" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.171:8774/\": EOF" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.093539 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.093728 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e8fe9843-7cb8-4d5f-9ff1-b2053d56666e" containerName="nova-metadata-log" containerID="cri-o://fac71ac72b0bd975678a5f8f609d60cba7908c936f1ef0df73f8d09495411ec0" gracePeriod=30 Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.094096 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e8fe9843-7cb8-4d5f-9ff1-b2053d56666e" containerName="nova-metadata-metadata" containerID="cri-o://87868c82af98dc65f5e0b08a41f6da75a7048153f3aa5c476f11215d1c792da6" gracePeriod=30 Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.211810 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.212067 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.359030 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.362544 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-z46bq" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.379272 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-g2j2k"] Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.419734 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-g2j2k"] Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.455155 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-config-data\") pod \"d4c1d967-acf9-4dbb-83db-1f654d4be6d2\" (UID: \"d4c1d967-acf9-4dbb-83db-1f654d4be6d2\") " Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.455264 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsm5g\" (UniqueName: \"kubernetes.io/projected/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-kube-api-access-dsm5g\") pod \"d4c1d967-acf9-4dbb-83db-1f654d4be6d2\" (UID: \"d4c1d967-acf9-4dbb-83db-1f654d4be6d2\") " Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.455434 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-combined-ca-bundle\") pod \"d4c1d967-acf9-4dbb-83db-1f654d4be6d2\" (UID: \"d4c1d967-acf9-4dbb-83db-1f654d4be6d2\") " Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.455464 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-scripts\") pod \"d4c1d967-acf9-4dbb-83db-1f654d4be6d2\" (UID: \"d4c1d967-acf9-4dbb-83db-1f654d4be6d2\") " Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.461471 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-scripts" (OuterVolumeSpecName: "scripts") pod "d4c1d967-acf9-4dbb-83db-1f654d4be6d2" (UID: "d4c1d967-acf9-4dbb-83db-1f654d4be6d2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.461560 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-kube-api-access-dsm5g" (OuterVolumeSpecName: "kube-api-access-dsm5g") pod "d4c1d967-acf9-4dbb-83db-1f654d4be6d2" (UID: "d4c1d967-acf9-4dbb-83db-1f654d4be6d2"). InnerVolumeSpecName "kube-api-access-dsm5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.491792 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-config-data" (OuterVolumeSpecName: "config-data") pod "d4c1d967-acf9-4dbb-83db-1f654d4be6d2" (UID: "d4c1d967-acf9-4dbb-83db-1f654d4be6d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.494276 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4c1d967-acf9-4dbb-83db-1f654d4be6d2" (UID: "d4c1d967-acf9-4dbb-83db-1f654d4be6d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.557533 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.557561 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.557570 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.557681 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsm5g\" (UniqueName: \"kubernetes.io/projected/d4c1d967-acf9-4dbb-83db-1f654d4be6d2-kube-api-access-dsm5g\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.609532 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.760641 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-combined-ca-bundle\") pod \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\" (UID: \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\") " Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.760991 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-config-data\") pod \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\" (UID: \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\") " Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.761049 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-logs\") pod \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\" (UID: \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\") " Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.761100 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4lpc\" (UniqueName: \"kubernetes.io/projected/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-kube-api-access-t4lpc\") pod \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\" (UID: \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\") " Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.761184 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-nova-metadata-tls-certs\") pod \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\" (UID: \"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e\") " Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.761341 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-logs" (OuterVolumeSpecName: "logs") pod "e8fe9843-7cb8-4d5f-9ff1-b2053d56666e" (UID: "e8fe9843-7cb8-4d5f-9ff1-b2053d56666e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.761575 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-logs\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.767466 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-kube-api-access-t4lpc" (OuterVolumeSpecName: "kube-api-access-t4lpc") pod "e8fe9843-7cb8-4d5f-9ff1-b2053d56666e" (UID: "e8fe9843-7cb8-4d5f-9ff1-b2053d56666e"). InnerVolumeSpecName "kube-api-access-t4lpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.785998 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-config-data" (OuterVolumeSpecName: "config-data") pod "e8fe9843-7cb8-4d5f-9ff1-b2053d56666e" (UID: "e8fe9843-7cb8-4d5f-9ff1-b2053d56666e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.794696 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8fe9843-7cb8-4d5f-9ff1-b2053d56666e" (UID: "e8fe9843-7cb8-4d5f-9ff1-b2053d56666e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.843081 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "e8fe9843-7cb8-4d5f-9ff1-b2053d56666e" (UID: "e8fe9843-7cb8-4d5f-9ff1-b2053d56666e"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.846386 4808 generic.go:334] "Generic (PLEG): container finished" podID="e8fe9843-7cb8-4d5f-9ff1-b2053d56666e" containerID="87868c82af98dc65f5e0b08a41f6da75a7048153f3aa5c476f11215d1c792da6" exitCode=0 Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.846422 4808 generic.go:334] "Generic (PLEG): container finished" podID="e8fe9843-7cb8-4d5f-9ff1-b2053d56666e" containerID="fac71ac72b0bd975678a5f8f609d60cba7908c936f1ef0df73f8d09495411ec0" exitCode=143 Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.846469 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e","Type":"ContainerDied","Data":"87868c82af98dc65f5e0b08a41f6da75a7048153f3aa5c476f11215d1c792da6"} Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.846497 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e","Type":"ContainerDied","Data":"fac71ac72b0bd975678a5f8f609d60cba7908c936f1ef0df73f8d09495411ec0"} Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.846511 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e8fe9843-7cb8-4d5f-9ff1-b2053d56666e","Type":"ContainerDied","Data":"8014d5ac162322b414aefcc3257685427d75432fe10755b72325b8d8ae1abb8c"} Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.846529 4808 scope.go:117] "RemoveContainer" containerID="87868c82af98dc65f5e0b08a41f6da75a7048153f3aa5c476f11215d1c792da6" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.846577 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.854795 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-z46bq" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.854783 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-z46bq" event={"ID":"d4c1d967-acf9-4dbb-83db-1f654d4be6d2","Type":"ContainerDied","Data":"fbdb3f7bec488856e35b21a8b6eb6ac008ee8f3ddd161a7e138e9932c45290d3"} Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.854931 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbdb3f7bec488856e35b21a8b6eb6ac008ee8f3ddd161a7e138e9932c45290d3" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.858157 4808 generic.go:334] "Generic (PLEG): container finished" podID="6c081cab-30c8-4952-8527-65355562b158" containerID="9978bc5c80a90a82c85e7d8655a1c7db8d94b3bd63746bf1c6fefae09a8a29b8" exitCode=143 Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.858770 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6c081cab-30c8-4952-8527-65355562b158","Type":"ContainerDied","Data":"9978bc5c80a90a82c85e7d8655a1c7db8d94b3bd63746bf1c6fefae09a8a29b8"} Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.872092 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4lpc\" (UniqueName: \"kubernetes.io/projected/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-kube-api-access-t4lpc\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.872116 4808 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.872127 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.872136 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.904269 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.905739 4808 scope.go:117] "RemoveContainer" containerID="fac71ac72b0bd975678a5f8f609d60cba7908c936f1ef0df73f8d09495411ec0" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.926319 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.930575 4808 scope.go:117] "RemoveContainer" containerID="87868c82af98dc65f5e0b08a41f6da75a7048153f3aa5c476f11215d1c792da6" Oct 02 16:59:55 crc kubenswrapper[4808]: E1002 16:59:55.938693 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87868c82af98dc65f5e0b08a41f6da75a7048153f3aa5c476f11215d1c792da6\": container with ID starting with 87868c82af98dc65f5e0b08a41f6da75a7048153f3aa5c476f11215d1c792da6 not found: ID does not exist" containerID="87868c82af98dc65f5e0b08a41f6da75a7048153f3aa5c476f11215d1c792da6" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.938736 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87868c82af98dc65f5e0b08a41f6da75a7048153f3aa5c476f11215d1c792da6"} err="failed to get container status \"87868c82af98dc65f5e0b08a41f6da75a7048153f3aa5c476f11215d1c792da6\": rpc error: code = NotFound desc = could not find container \"87868c82af98dc65f5e0b08a41f6da75a7048153f3aa5c476f11215d1c792da6\": container with ID starting with 87868c82af98dc65f5e0b08a41f6da75a7048153f3aa5c476f11215d1c792da6 not found: ID does not exist" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.938760 4808 scope.go:117] "RemoveContainer" containerID="fac71ac72b0bd975678a5f8f609d60cba7908c936f1ef0df73f8d09495411ec0" Oct 02 16:59:55 crc kubenswrapper[4808]: E1002 16:59:55.940033 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fac71ac72b0bd975678a5f8f609d60cba7908c936f1ef0df73f8d09495411ec0\": container with ID starting with fac71ac72b0bd975678a5f8f609d60cba7908c936f1ef0df73f8d09495411ec0 not found: ID does not exist" containerID="fac71ac72b0bd975678a5f8f609d60cba7908c936f1ef0df73f8d09495411ec0" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.940066 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fac71ac72b0bd975678a5f8f609d60cba7908c936f1ef0df73f8d09495411ec0"} err="failed to get container status \"fac71ac72b0bd975678a5f8f609d60cba7908c936f1ef0df73f8d09495411ec0\": rpc error: code = NotFound desc = could not find container \"fac71ac72b0bd975678a5f8f609d60cba7908c936f1ef0df73f8d09495411ec0\": container with ID starting with fac71ac72b0bd975678a5f8f609d60cba7908c936f1ef0df73f8d09495411ec0 not found: ID does not exist" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.940081 4808 scope.go:117] "RemoveContainer" containerID="87868c82af98dc65f5e0b08a41f6da75a7048153f3aa5c476f11215d1c792da6" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.942557 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.942781 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87868c82af98dc65f5e0b08a41f6da75a7048153f3aa5c476f11215d1c792da6"} err="failed to get container status \"87868c82af98dc65f5e0b08a41f6da75a7048153f3aa5c476f11215d1c792da6\": rpc error: code = NotFound desc = could not find container \"87868c82af98dc65f5e0b08a41f6da75a7048153f3aa5c476f11215d1c792da6\": container with ID starting with 87868c82af98dc65f5e0b08a41f6da75a7048153f3aa5c476f11215d1c792da6 not found: ID does not exist" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.942854 4808 scope.go:117] "RemoveContainer" containerID="fac71ac72b0bd975678a5f8f609d60cba7908c936f1ef0df73f8d09495411ec0" Oct 02 16:59:55 crc kubenswrapper[4808]: E1002 16:59:55.943106 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c166b259-1f16-4e18-8df1-a4a222d49897" containerName="dnsmasq-dns" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.943133 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c166b259-1f16-4e18-8df1-a4a222d49897" containerName="dnsmasq-dns" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.943142 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fac71ac72b0bd975678a5f8f609d60cba7908c936f1ef0df73f8d09495411ec0"} err="failed to get container status \"fac71ac72b0bd975678a5f8f609d60cba7908c936f1ef0df73f8d09495411ec0\": rpc error: code = NotFound desc = could not find container \"fac71ac72b0bd975678a5f8f609d60cba7908c936f1ef0df73f8d09495411ec0\": container with ID starting with fac71ac72b0bd975678a5f8f609d60cba7908c936f1ef0df73f8d09495411ec0 not found: ID does not exist" Oct 02 16:59:55 crc kubenswrapper[4808]: E1002 16:59:55.943150 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4c1d967-acf9-4dbb-83db-1f654d4be6d2" containerName="nova-cell1-conductor-db-sync" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.943170 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4c1d967-acf9-4dbb-83db-1f654d4be6d2" containerName="nova-cell1-conductor-db-sync" Oct 02 16:59:55 crc kubenswrapper[4808]: E1002 16:59:55.943224 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4bb01f7-7c71-4b46-b45d-49d3cd6b0958" containerName="nova-manage" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.943259 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4bb01f7-7c71-4b46-b45d-49d3cd6b0958" containerName="nova-manage" Oct 02 16:59:55 crc kubenswrapper[4808]: E1002 16:59:55.943270 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8fe9843-7cb8-4d5f-9ff1-b2053d56666e" containerName="nova-metadata-metadata" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.943278 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8fe9843-7cb8-4d5f-9ff1-b2053d56666e" containerName="nova-metadata-metadata" Oct 02 16:59:55 crc kubenswrapper[4808]: E1002 16:59:55.943288 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c166b259-1f16-4e18-8df1-a4a222d49897" containerName="init" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.943294 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c166b259-1f16-4e18-8df1-a4a222d49897" containerName="init" Oct 02 16:59:55 crc kubenswrapper[4808]: E1002 16:59:55.943322 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8fe9843-7cb8-4d5f-9ff1-b2053d56666e" containerName="nova-metadata-log" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.943329 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8fe9843-7cb8-4d5f-9ff1-b2053d56666e" containerName="nova-metadata-log" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.943583 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4c1d967-acf9-4dbb-83db-1f654d4be6d2" containerName="nova-cell1-conductor-db-sync" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.943603 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8fe9843-7cb8-4d5f-9ff1-b2053d56666e" containerName="nova-metadata-log" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.943616 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8fe9843-7cb8-4d5f-9ff1-b2053d56666e" containerName="nova-metadata-metadata" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.943627 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4bb01f7-7c71-4b46-b45d-49d3cd6b0958" containerName="nova-manage" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.943634 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c166b259-1f16-4e18-8df1-a4a222d49897" containerName="dnsmasq-dns" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.944210 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.947604 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.958563 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.967389 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.973056 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.973755 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.974902 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 16:59:55 crc kubenswrapper[4808]: I1002 16:59:55.978891 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.076615 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqtrf\" (UniqueName: \"kubernetes.io/projected/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-kube-api-access-vqtrf\") pod \"nova-metadata-0\" (UID: \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\") " pod="openstack/nova-metadata-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.076763 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s84zq\" (UniqueName: \"kubernetes.io/projected/75314254-401d-44e1-b68b-658aad79a9df-kube-api-access-s84zq\") pod \"nova-cell1-conductor-0\" (UID: \"75314254-401d-44e1-b68b-658aad79a9df\") " pod="openstack/nova-cell1-conductor-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.076818 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-logs\") pod \"nova-metadata-0\" (UID: \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\") " pod="openstack/nova-metadata-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.076872 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75314254-401d-44e1-b68b-658aad79a9df-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"75314254-401d-44e1-b68b-658aad79a9df\") " pod="openstack/nova-cell1-conductor-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.076952 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-config-data\") pod \"nova-metadata-0\" (UID: \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\") " pod="openstack/nova-metadata-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.076974 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\") " pod="openstack/nova-metadata-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.077093 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\") " pod="openstack/nova-metadata-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.077130 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75314254-401d-44e1-b68b-658aad79a9df-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"75314254-401d-44e1-b68b-658aad79a9df\") " pod="openstack/nova-cell1-conductor-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.178295 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\") " pod="openstack/nova-metadata-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.178391 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75314254-401d-44e1-b68b-658aad79a9df-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"75314254-401d-44e1-b68b-658aad79a9df\") " pod="openstack/nova-cell1-conductor-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.178579 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqtrf\" (UniqueName: \"kubernetes.io/projected/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-kube-api-access-vqtrf\") pod \"nova-metadata-0\" (UID: \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\") " pod="openstack/nova-metadata-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.178718 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s84zq\" (UniqueName: \"kubernetes.io/projected/75314254-401d-44e1-b68b-658aad79a9df-kube-api-access-s84zq\") pod \"nova-cell1-conductor-0\" (UID: \"75314254-401d-44e1-b68b-658aad79a9df\") " pod="openstack/nova-cell1-conductor-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.178794 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-logs\") pod \"nova-metadata-0\" (UID: \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\") " pod="openstack/nova-metadata-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.178837 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75314254-401d-44e1-b68b-658aad79a9df-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"75314254-401d-44e1-b68b-658aad79a9df\") " pod="openstack/nova-cell1-conductor-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.178945 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-config-data\") pod \"nova-metadata-0\" (UID: \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\") " pod="openstack/nova-metadata-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.178973 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\") " pod="openstack/nova-metadata-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.179105 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-logs\") pod \"nova-metadata-0\" (UID: \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\") " pod="openstack/nova-metadata-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.183874 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75314254-401d-44e1-b68b-658aad79a9df-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"75314254-401d-44e1-b68b-658aad79a9df\") " pod="openstack/nova-cell1-conductor-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.184820 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75314254-401d-44e1-b68b-658aad79a9df-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"75314254-401d-44e1-b68b-658aad79a9df\") " pod="openstack/nova-cell1-conductor-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.190946 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-config-data\") pod \"nova-metadata-0\" (UID: \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\") " pod="openstack/nova-metadata-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.195702 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\") " pod="openstack/nova-metadata-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.195873 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\") " pod="openstack/nova-metadata-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.197558 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s84zq\" (UniqueName: \"kubernetes.io/projected/75314254-401d-44e1-b68b-658aad79a9df-kube-api-access-s84zq\") pod \"nova-cell1-conductor-0\" (UID: \"75314254-401d-44e1-b68b-658aad79a9df\") " pod="openstack/nova-cell1-conductor-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.201746 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqtrf\" (UniqueName: \"kubernetes.io/projected/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-kube-api-access-vqtrf\") pod \"nova-metadata-0\" (UID: \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\") " pod="openstack/nova-metadata-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.265129 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.298508 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.789856 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.866310 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d7a4dd0-7af5-4d01-b888-14ce54434b2d","Type":"ContainerStarted","Data":"68650fb34154384ab162335c1acb367c0475844cbefa9e8eca63fd27af9aec52"} Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.868505 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="b4b16f51-ca0e-41eb-9033-86a5f7ee6091" containerName="nova-scheduler-scheduler" containerID="cri-o://96d3a2743013e5b72aaab1a8e1bf4ba1d55a8160eabc25b4cdf5262c3a5c0fba" gracePeriod=30 Oct 02 16:59:56 crc kubenswrapper[4808]: I1002 16:59:56.869317 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 16:59:57 crc kubenswrapper[4808]: I1002 16:59:57.419125 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c166b259-1f16-4e18-8df1-a4a222d49897" path="/var/lib/kubelet/pods/c166b259-1f16-4e18-8df1-a4a222d49897/volumes" Oct 02 16:59:57 crc kubenswrapper[4808]: I1002 16:59:57.419972 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8fe9843-7cb8-4d5f-9ff1-b2053d56666e" path="/var/lib/kubelet/pods/e8fe9843-7cb8-4d5f-9ff1-b2053d56666e/volumes" Oct 02 16:59:57 crc kubenswrapper[4808]: I1002 16:59:57.882471 4808 generic.go:334] "Generic (PLEG): container finished" podID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" containerID="3d0dd557125f47597903502f7aea47322f8712e52bf3c8140a68ed8715fe94da" exitCode=0 Oct 02 16:59:57 crc kubenswrapper[4808]: I1002 16:59:57.882527 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7","Type":"ContainerDied","Data":"3d0dd557125f47597903502f7aea47322f8712e52bf3c8140a68ed8715fe94da"} Oct 02 16:59:57 crc kubenswrapper[4808]: I1002 16:59:57.887444 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d7a4dd0-7af5-4d01-b888-14ce54434b2d","Type":"ContainerStarted","Data":"41cb07534ca1da05f1b1ba049073f435695baf47a1d14f6e92fbba1ee4be4032"} Oct 02 16:59:57 crc kubenswrapper[4808]: I1002 16:59:57.887486 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d7a4dd0-7af5-4d01-b888-14ce54434b2d","Type":"ContainerStarted","Data":"f49fb6be9af659916b2b1aa693d686204de7fd4bfc4f8c347de9f151a61c84d8"} Oct 02 16:59:57 crc kubenswrapper[4808]: I1002 16:59:57.892555 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"75314254-401d-44e1-b68b-658aad79a9df","Type":"ContainerStarted","Data":"fb6c261e7c96bce6cea377a107b19fff7155b12676007ee5ba1e856e3e191af9"} Oct 02 16:59:57 crc kubenswrapper[4808]: I1002 16:59:57.892587 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"75314254-401d-44e1-b68b-658aad79a9df","Type":"ContainerStarted","Data":"ee5a5beba2e2cb4942df88c7d97c89d00acbf9649c7f1fa2dbf659e098dc845a"} Oct 02 16:59:57 crc kubenswrapper[4808]: I1002 16:59:57.892681 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 02 16:59:57 crc kubenswrapper[4808]: I1002 16:59:57.916964 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.916945642 podStartE2EDuration="2.916945642s" podCreationTimestamp="2025-10-02 16:59:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:59:57.909364396 +0000 UTC m=+1165.234893416" watchObservedRunningTime="2025-10-02 16:59:57.916945642 +0000 UTC m=+1165.242474652" Oct 02 16:59:57 crc kubenswrapper[4808]: I1002 16:59:57.934825 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.934800256 podStartE2EDuration="2.934800256s" podCreationTimestamp="2025-10-02 16:59:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 16:59:57.92978154 +0000 UTC m=+1165.255310550" watchObservedRunningTime="2025-10-02 16:59:57.934800256 +0000 UTC m=+1165.260329266" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.026753 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.129889 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-combined-ca-bundle\") pod \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.129974 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-config-data\") pod \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.130000 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-sg-core-conf-yaml\") pod \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.130122 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8kmr\" (UniqueName: \"kubernetes.io/projected/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-kube-api-access-w8kmr\") pod \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.130160 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-log-httpd\") pod \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.130223 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-scripts\") pod \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.130299 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-run-httpd\") pod \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\" (UID: \"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7\") " Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.131391 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" (UID: "596a4d5a-b4ba-466b-a9e5-ac4c853efcb7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.131732 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" (UID: "596a4d5a-b4ba-466b-a9e5-ac4c853efcb7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.137448 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-kube-api-access-w8kmr" (OuterVolumeSpecName: "kube-api-access-w8kmr") pod "596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" (UID: "596a4d5a-b4ba-466b-a9e5-ac4c853efcb7"). InnerVolumeSpecName "kube-api-access-w8kmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.138999 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-scripts" (OuterVolumeSpecName: "scripts") pod "596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" (UID: "596a4d5a-b4ba-466b-a9e5-ac4c853efcb7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.165967 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" (UID: "596a4d5a-b4ba-466b-a9e5-ac4c853efcb7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.213461 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" (UID: "596a4d5a-b4ba-466b-a9e5-ac4c853efcb7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.232644 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8kmr\" (UniqueName: \"kubernetes.io/projected/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-kube-api-access-w8kmr\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.232669 4808 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.232679 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.232686 4808 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.232694 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.232701 4808 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.236452 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-config-data" (OuterVolumeSpecName: "config-data") pod "596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" (UID: "596a4d5a-b4ba-466b-a9e5-ac4c853efcb7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.334867 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.908724 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"596a4d5a-b4ba-466b-a9e5-ac4c853efcb7","Type":"ContainerDied","Data":"cce32853ea1e26a3b1a35a26b1e8de066a35b21f3d7ac227030f2ad07a363b55"} Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.908821 4808 scope.go:117] "RemoveContainer" containerID="fe7ced19e6b556a1b7a881eaf296ab31a77bc3a70e065fde7fd3dd34a1714fe9" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.909272 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.958962 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.970414 4808 scope.go:117] "RemoveContainer" containerID="94522ef9e646b66e17dc4bfa2f54fe131a24052e954ad1d679f5b93845992bf0" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.974539 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.999391 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:59:58 crc kubenswrapper[4808]: E1002 16:59:58.999801 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" containerName="sg-core" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.999825 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" containerName="sg-core" Oct 02 16:59:58 crc kubenswrapper[4808]: E1002 16:59:58.999855 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" containerName="ceilometer-notification-agent" Oct 02 16:59:58 crc kubenswrapper[4808]: I1002 16:59:58.999865 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" containerName="ceilometer-notification-agent" Oct 02 16:59:58 crc kubenswrapper[4808]: E1002 16:59:58.999890 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" containerName="ceilometer-central-agent" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:58.999899 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" containerName="ceilometer-central-agent" Oct 02 16:59:59 crc kubenswrapper[4808]: E1002 16:59:58.999910 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" containerName="proxy-httpd" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:58.999918 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" containerName="proxy-httpd" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.000146 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" containerName="ceilometer-notification-agent" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.000177 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" containerName="ceilometer-central-agent" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.000196 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" containerName="proxy-httpd" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.000224 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" containerName="sg-core" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.004658 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.009066 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.015749 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.015804 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.016062 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.027895 4808 scope.go:117] "RemoveContainer" containerID="3d0dd557125f47597903502f7aea47322f8712e52bf3c8140a68ed8715fe94da" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.056043 4808 scope.go:117] "RemoveContainer" containerID="80ed9069762936ea68104c7c75c37bfd9b082fa028cd3d54e34778db6bf53a6a" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.155508 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c66f29c7-9e93-4417-bc3e-b486a0b167cc-run-httpd\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.155774 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-config-data\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.155905 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.156019 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.156121 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.156221 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcnsx\" (UniqueName: \"kubernetes.io/projected/c66f29c7-9e93-4417-bc3e-b486a0b167cc-kube-api-access-wcnsx\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.156499 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-scripts\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.156626 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c66f29c7-9e93-4417-bc3e-b486a0b167cc-log-httpd\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.258393 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.258461 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.258520 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.258569 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcnsx\" (UniqueName: \"kubernetes.io/projected/c66f29c7-9e93-4417-bc3e-b486a0b167cc-kube-api-access-wcnsx\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.258649 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-scripts\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.258683 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c66f29c7-9e93-4417-bc3e-b486a0b167cc-log-httpd\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.258792 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c66f29c7-9e93-4417-bc3e-b486a0b167cc-run-httpd\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.258827 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-config-data\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.259594 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c66f29c7-9e93-4417-bc3e-b486a0b167cc-log-httpd\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.259644 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c66f29c7-9e93-4417-bc3e-b486a0b167cc-run-httpd\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.263503 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.265199 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.274589 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcnsx\" (UniqueName: \"kubernetes.io/projected/c66f29c7-9e93-4417-bc3e-b486a0b167cc-kube-api-access-wcnsx\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.274609 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-config-data\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.276616 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.279527 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-scripts\") pod \"ceilometer-0\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: E1002 16:59:59.310541 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="96d3a2743013e5b72aaab1a8e1bf4ba1d55a8160eabc25b4cdf5262c3a5c0fba" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 16:59:59 crc kubenswrapper[4808]: E1002 16:59:59.312194 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="96d3a2743013e5b72aaab1a8e1bf4ba1d55a8160eabc25b4cdf5262c3a5c0fba" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 16:59:59 crc kubenswrapper[4808]: E1002 16:59:59.316107 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="96d3a2743013e5b72aaab1a8e1bf4ba1d55a8160eabc25b4cdf5262c3a5c0fba" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 16:59:59 crc kubenswrapper[4808]: E1002 16:59:59.316166 4808 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="b4b16f51-ca0e-41eb-9033-86a5f7ee6091" containerName="nova-scheduler-scheduler" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.335625 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.410206 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="596a4d5a-b4ba-466b-a9e5-ac4c853efcb7" path="/var/lib/kubelet/pods/596a4d5a-b4ba-466b-a9e5-ac4c853efcb7/volumes" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.784275 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 16:59:59 crc kubenswrapper[4808]: W1002 16:59:59.805648 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc66f29c7_9e93_4417_bc3e_b486a0b167cc.slice/crio-76e577573bdd504776f07aac31b5ff3c20243b4ddb7d01494f3dcc9e7992e2d9 WatchSource:0}: Error finding container 76e577573bdd504776f07aac31b5ff3c20243b4ddb7d01494f3dcc9e7992e2d9: Status 404 returned error can't find the container with id 76e577573bdd504776f07aac31b5ff3c20243b4ddb7d01494f3dcc9e7992e2d9 Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.824480 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.920624 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c66f29c7-9e93-4417-bc3e-b486a0b167cc","Type":"ContainerStarted","Data":"76e577573bdd504776f07aac31b5ff3c20243b4ddb7d01494f3dcc9e7992e2d9"} Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.925050 4808 generic.go:334] "Generic (PLEG): container finished" podID="b4b16f51-ca0e-41eb-9033-86a5f7ee6091" containerID="96d3a2743013e5b72aaab1a8e1bf4ba1d55a8160eabc25b4cdf5262c3a5c0fba" exitCode=0 Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.925085 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b4b16f51-ca0e-41eb-9033-86a5f7ee6091","Type":"ContainerDied","Data":"96d3a2743013e5b72aaab1a8e1bf4ba1d55a8160eabc25b4cdf5262c3a5c0fba"} Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.925106 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b4b16f51-ca0e-41eb-9033-86a5f7ee6091","Type":"ContainerDied","Data":"9ccfe9d28bd411d3138d5f3e9e40ff85060232033069ac6edf9d314714f03ca8"} Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.925125 4808 scope.go:117] "RemoveContainer" containerID="96d3a2743013e5b72aaab1a8e1bf4ba1d55a8160eabc25b4cdf5262c3a5c0fba" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.925271 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.953251 4808 scope.go:117] "RemoveContainer" containerID="96d3a2743013e5b72aaab1a8e1bf4ba1d55a8160eabc25b4cdf5262c3a5c0fba" Oct 02 16:59:59 crc kubenswrapper[4808]: E1002 16:59:59.954557 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96d3a2743013e5b72aaab1a8e1bf4ba1d55a8160eabc25b4cdf5262c3a5c0fba\": container with ID starting with 96d3a2743013e5b72aaab1a8e1bf4ba1d55a8160eabc25b4cdf5262c3a5c0fba not found: ID does not exist" containerID="96d3a2743013e5b72aaab1a8e1bf4ba1d55a8160eabc25b4cdf5262c3a5c0fba" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.954616 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96d3a2743013e5b72aaab1a8e1bf4ba1d55a8160eabc25b4cdf5262c3a5c0fba"} err="failed to get container status \"96d3a2743013e5b72aaab1a8e1bf4ba1d55a8160eabc25b4cdf5262c3a5c0fba\": rpc error: code = NotFound desc = could not find container \"96d3a2743013e5b72aaab1a8e1bf4ba1d55a8160eabc25b4cdf5262c3a5c0fba\": container with ID starting with 96d3a2743013e5b72aaab1a8e1bf4ba1d55a8160eabc25b4cdf5262c3a5c0fba not found: ID does not exist" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.974271 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4b16f51-ca0e-41eb-9033-86a5f7ee6091-combined-ca-bundle\") pod \"b4b16f51-ca0e-41eb-9033-86a5f7ee6091\" (UID: \"b4b16f51-ca0e-41eb-9033-86a5f7ee6091\") " Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.974329 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rkvc\" (UniqueName: \"kubernetes.io/projected/b4b16f51-ca0e-41eb-9033-86a5f7ee6091-kube-api-access-4rkvc\") pod \"b4b16f51-ca0e-41eb-9033-86a5f7ee6091\" (UID: \"b4b16f51-ca0e-41eb-9033-86a5f7ee6091\") " Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.974432 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4b16f51-ca0e-41eb-9033-86a5f7ee6091-config-data\") pod \"b4b16f51-ca0e-41eb-9033-86a5f7ee6091\" (UID: \"b4b16f51-ca0e-41eb-9033-86a5f7ee6091\") " Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.979536 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4b16f51-ca0e-41eb-9033-86a5f7ee6091-kube-api-access-4rkvc" (OuterVolumeSpecName: "kube-api-access-4rkvc") pod "b4b16f51-ca0e-41eb-9033-86a5f7ee6091" (UID: "b4b16f51-ca0e-41eb-9033-86a5f7ee6091"). InnerVolumeSpecName "kube-api-access-4rkvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 16:59:59 crc kubenswrapper[4808]: I1002 16:59:59.998969 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4b16f51-ca0e-41eb-9033-86a5f7ee6091-config-data" (OuterVolumeSpecName: "config-data") pod "b4b16f51-ca0e-41eb-9033-86a5f7ee6091" (UID: "b4b16f51-ca0e-41eb-9033-86a5f7ee6091"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.024857 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4b16f51-ca0e-41eb-9033-86a5f7ee6091-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b4b16f51-ca0e-41eb-9033-86a5f7ee6091" (UID: "b4b16f51-ca0e-41eb-9033-86a5f7ee6091"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.076662 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rkvc\" (UniqueName: \"kubernetes.io/projected/b4b16f51-ca0e-41eb-9033-86a5f7ee6091-kube-api-access-4rkvc\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.076689 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4b16f51-ca0e-41eb-9033-86a5f7ee6091-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.076702 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4b16f51-ca0e-41eb-9033-86a5f7ee6091-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.139064 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323740-2zgrw"] Oct 02 17:00:00 crc kubenswrapper[4808]: E1002 17:00:00.139418 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b16f51-ca0e-41eb-9033-86a5f7ee6091" containerName="nova-scheduler-scheduler" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.139434 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b16f51-ca0e-41eb-9033-86a5f7ee6091" containerName="nova-scheduler-scheduler" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.139603 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4b16f51-ca0e-41eb-9033-86a5f7ee6091" containerName="nova-scheduler-scheduler" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.140144 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323740-2zgrw" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.142915 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.143302 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.151064 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323740-2zgrw"] Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.259263 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.266683 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.279577 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.280713 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvpf4\" (UniqueName: \"kubernetes.io/projected/c15a4b91-873b-429f-9b0c-306cda038d5e-kube-api-access-gvpf4\") pod \"collect-profiles-29323740-2zgrw\" (UID: \"c15a4b91-873b-429f-9b0c-306cda038d5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323740-2zgrw" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.280934 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.281053 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c15a4b91-873b-429f-9b0c-306cda038d5e-config-volume\") pod \"collect-profiles-29323740-2zgrw\" (UID: \"c15a4b91-873b-429f-9b0c-306cda038d5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323740-2zgrw" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.281308 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c15a4b91-873b-429f-9b0c-306cda038d5e-secret-volume\") pod \"collect-profiles-29323740-2zgrw\" (UID: \"c15a4b91-873b-429f-9b0c-306cda038d5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323740-2zgrw" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.283490 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.291171 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.382638 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmrh4\" (UniqueName: \"kubernetes.io/projected/4aa280f6-f14f-47de-bbba-8f49229e4ec4-kube-api-access-zmrh4\") pod \"nova-scheduler-0\" (UID: \"4aa280f6-f14f-47de-bbba-8f49229e4ec4\") " pod="openstack/nova-scheduler-0" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.382829 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c15a4b91-873b-429f-9b0c-306cda038d5e-secret-volume\") pod \"collect-profiles-29323740-2zgrw\" (UID: \"c15a4b91-873b-429f-9b0c-306cda038d5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323740-2zgrw" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.382886 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvpf4\" (UniqueName: \"kubernetes.io/projected/c15a4b91-873b-429f-9b0c-306cda038d5e-kube-api-access-gvpf4\") pod \"collect-profiles-29323740-2zgrw\" (UID: \"c15a4b91-873b-429f-9b0c-306cda038d5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323740-2zgrw" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.382941 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c15a4b91-873b-429f-9b0c-306cda038d5e-config-volume\") pod \"collect-profiles-29323740-2zgrw\" (UID: \"c15a4b91-873b-429f-9b0c-306cda038d5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323740-2zgrw" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.382966 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa280f6-f14f-47de-bbba-8f49229e4ec4-config-data\") pod \"nova-scheduler-0\" (UID: \"4aa280f6-f14f-47de-bbba-8f49229e4ec4\") " pod="openstack/nova-scheduler-0" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.383004 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa280f6-f14f-47de-bbba-8f49229e4ec4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4aa280f6-f14f-47de-bbba-8f49229e4ec4\") " pod="openstack/nova-scheduler-0" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.385090 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c15a4b91-873b-429f-9b0c-306cda038d5e-config-volume\") pod \"collect-profiles-29323740-2zgrw\" (UID: \"c15a4b91-873b-429f-9b0c-306cda038d5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323740-2zgrw" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.389338 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c15a4b91-873b-429f-9b0c-306cda038d5e-secret-volume\") pod \"collect-profiles-29323740-2zgrw\" (UID: \"c15a4b91-873b-429f-9b0c-306cda038d5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323740-2zgrw" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.410703 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvpf4\" (UniqueName: \"kubernetes.io/projected/c15a4b91-873b-429f-9b0c-306cda038d5e-kube-api-access-gvpf4\") pod \"collect-profiles-29323740-2zgrw\" (UID: \"c15a4b91-873b-429f-9b0c-306cda038d5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323740-2zgrw" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.460760 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323740-2zgrw" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.486073 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa280f6-f14f-47de-bbba-8f49229e4ec4-config-data\") pod \"nova-scheduler-0\" (UID: \"4aa280f6-f14f-47de-bbba-8f49229e4ec4\") " pod="openstack/nova-scheduler-0" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.486413 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa280f6-f14f-47de-bbba-8f49229e4ec4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4aa280f6-f14f-47de-bbba-8f49229e4ec4\") " pod="openstack/nova-scheduler-0" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.486510 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmrh4\" (UniqueName: \"kubernetes.io/projected/4aa280f6-f14f-47de-bbba-8f49229e4ec4-kube-api-access-zmrh4\") pod \"nova-scheduler-0\" (UID: \"4aa280f6-f14f-47de-bbba-8f49229e4ec4\") " pod="openstack/nova-scheduler-0" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.492814 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa280f6-f14f-47de-bbba-8f49229e4ec4-config-data\") pod \"nova-scheduler-0\" (UID: \"4aa280f6-f14f-47de-bbba-8f49229e4ec4\") " pod="openstack/nova-scheduler-0" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.501020 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa280f6-f14f-47de-bbba-8f49229e4ec4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4aa280f6-f14f-47de-bbba-8f49229e4ec4\") " pod="openstack/nova-scheduler-0" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.513605 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmrh4\" (UniqueName: \"kubernetes.io/projected/4aa280f6-f14f-47de-bbba-8f49229e4ec4-kube-api-access-zmrh4\") pod \"nova-scheduler-0\" (UID: \"4aa280f6-f14f-47de-bbba-8f49229e4ec4\") " pod="openstack/nova-scheduler-0" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.633030 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.702944 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.704102 4808 scope.go:117] "RemoveContainer" containerID="a918ee069243e5dc0b7673d53697e23876162204008e5ae0c51065fa9012a203" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.904710 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.905081 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323740-2zgrw"] Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.947689 4808 generic.go:334] "Generic (PLEG): container finished" podID="6c081cab-30c8-4952-8527-65355562b158" containerID="cafa5d95dfc4937e0adc6b08c59a5d0654e48df9151bc9ffb7dc1eeafb0c2dba" exitCode=0 Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.947959 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6c081cab-30c8-4952-8527-65355562b158","Type":"ContainerDied","Data":"cafa5d95dfc4937e0adc6b08c59a5d0654e48df9151bc9ffb7dc1eeafb0c2dba"} Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.947985 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6c081cab-30c8-4952-8527-65355562b158","Type":"ContainerDied","Data":"5382cf4df4e928ef18d9ff4685f5fc3678b23786128e30c7d109fa51200d34b4"} Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.948001 4808 scope.go:117] "RemoveContainer" containerID="cafa5d95dfc4937e0adc6b08c59a5d0654e48df9151bc9ffb7dc1eeafb0c2dba" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.948122 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.954155 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" event={"ID":"818e02a3-653a-4b36-a68b-9d4e2df0b71e","Type":"ContainerStarted","Data":"dc9639f10a3923cdab18975fdbea062b709d111e09fdf3caa99a2b5ab5ce0a27"} Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.954344 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.963189 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c66f29c7-9e93-4417-bc3e-b486a0b167cc","Type":"ContainerStarted","Data":"0f6479ff535d02280faff7cca75dffddfd4ae0daef90502722fe17ad3ffaa3fb"} Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.989351 4808 scope.go:117] "RemoveContainer" containerID="9978bc5c80a90a82c85e7d8655a1c7db8d94b3bd63746bf1c6fefae09a8a29b8" Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.996873 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c081cab-30c8-4952-8527-65355562b158-combined-ca-bundle\") pod \"6c081cab-30c8-4952-8527-65355562b158\" (UID: \"6c081cab-30c8-4952-8527-65355562b158\") " Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.996967 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jt682\" (UniqueName: \"kubernetes.io/projected/6c081cab-30c8-4952-8527-65355562b158-kube-api-access-jt682\") pod \"6c081cab-30c8-4952-8527-65355562b158\" (UID: \"6c081cab-30c8-4952-8527-65355562b158\") " Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.997049 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c081cab-30c8-4952-8527-65355562b158-config-data\") pod \"6c081cab-30c8-4952-8527-65355562b158\" (UID: \"6c081cab-30c8-4952-8527-65355562b158\") " Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.997115 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c081cab-30c8-4952-8527-65355562b158-logs\") pod \"6c081cab-30c8-4952-8527-65355562b158\" (UID: \"6c081cab-30c8-4952-8527-65355562b158\") " Oct 02 17:00:00 crc kubenswrapper[4808]: I1002 17:00:00.998314 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c081cab-30c8-4952-8527-65355562b158-logs" (OuterVolumeSpecName: "logs") pod "6c081cab-30c8-4952-8527-65355562b158" (UID: "6c081cab-30c8-4952-8527-65355562b158"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.001731 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c081cab-30c8-4952-8527-65355562b158-kube-api-access-jt682" (OuterVolumeSpecName: "kube-api-access-jt682") pod "6c081cab-30c8-4952-8527-65355562b158" (UID: "6c081cab-30c8-4952-8527-65355562b158"). InnerVolumeSpecName "kube-api-access-jt682". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.009086 4808 scope.go:117] "RemoveContainer" containerID="cafa5d95dfc4937e0adc6b08c59a5d0654e48df9151bc9ffb7dc1eeafb0c2dba" Oct 02 17:00:01 crc kubenswrapper[4808]: E1002 17:00:01.009584 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cafa5d95dfc4937e0adc6b08c59a5d0654e48df9151bc9ffb7dc1eeafb0c2dba\": container with ID starting with cafa5d95dfc4937e0adc6b08c59a5d0654e48df9151bc9ffb7dc1eeafb0c2dba not found: ID does not exist" containerID="cafa5d95dfc4937e0adc6b08c59a5d0654e48df9151bc9ffb7dc1eeafb0c2dba" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.009635 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cafa5d95dfc4937e0adc6b08c59a5d0654e48df9151bc9ffb7dc1eeafb0c2dba"} err="failed to get container status \"cafa5d95dfc4937e0adc6b08c59a5d0654e48df9151bc9ffb7dc1eeafb0c2dba\": rpc error: code = NotFound desc = could not find container \"cafa5d95dfc4937e0adc6b08c59a5d0654e48df9151bc9ffb7dc1eeafb0c2dba\": container with ID starting with cafa5d95dfc4937e0adc6b08c59a5d0654e48df9151bc9ffb7dc1eeafb0c2dba not found: ID does not exist" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.009668 4808 scope.go:117] "RemoveContainer" containerID="9978bc5c80a90a82c85e7d8655a1c7db8d94b3bd63746bf1c6fefae09a8a29b8" Oct 02 17:00:01 crc kubenswrapper[4808]: E1002 17:00:01.009937 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9978bc5c80a90a82c85e7d8655a1c7db8d94b3bd63746bf1c6fefae09a8a29b8\": container with ID starting with 9978bc5c80a90a82c85e7d8655a1c7db8d94b3bd63746bf1c6fefae09a8a29b8 not found: ID does not exist" containerID="9978bc5c80a90a82c85e7d8655a1c7db8d94b3bd63746bf1c6fefae09a8a29b8" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.009982 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9978bc5c80a90a82c85e7d8655a1c7db8d94b3bd63746bf1c6fefae09a8a29b8"} err="failed to get container status \"9978bc5c80a90a82c85e7d8655a1c7db8d94b3bd63746bf1c6fefae09a8a29b8\": rpc error: code = NotFound desc = could not find container \"9978bc5c80a90a82c85e7d8655a1c7db8d94b3bd63746bf1c6fefae09a8a29b8\": container with ID starting with 9978bc5c80a90a82c85e7d8655a1c7db8d94b3bd63746bf1c6fefae09a8a29b8 not found: ID does not exist" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.027199 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c081cab-30c8-4952-8527-65355562b158-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c081cab-30c8-4952-8527-65355562b158" (UID: "6c081cab-30c8-4952-8527-65355562b158"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.037217 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c081cab-30c8-4952-8527-65355562b158-config-data" (OuterVolumeSpecName: "config-data") pod "6c081cab-30c8-4952-8527-65355562b158" (UID: "6c081cab-30c8-4952-8527-65355562b158"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.101623 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c081cab-30c8-4952-8527-65355562b158-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.101658 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c081cab-30c8-4952-8527-65355562b158-logs\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.101669 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c081cab-30c8-4952-8527-65355562b158-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.101683 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jt682\" (UniqueName: \"kubernetes.io/projected/6c081cab-30c8-4952-8527-65355562b158-kube-api-access-jt682\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.111931 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.298834 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.299081 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.408449 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4b16f51-ca0e-41eb-9033-86a5f7ee6091" path="/var/lib/kubelet/pods/b4b16f51-ca0e-41eb-9033-86a5f7ee6091/volumes" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.409300 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.409325 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.426972 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 17:00:01 crc kubenswrapper[4808]: E1002 17:00:01.427457 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c081cab-30c8-4952-8527-65355562b158" containerName="nova-api-log" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.427472 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c081cab-30c8-4952-8527-65355562b158" containerName="nova-api-log" Oct 02 17:00:01 crc kubenswrapper[4808]: E1002 17:00:01.427493 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c081cab-30c8-4952-8527-65355562b158" containerName="nova-api-api" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.427501 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c081cab-30c8-4952-8527-65355562b158" containerName="nova-api-api" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.427682 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c081cab-30c8-4952-8527-65355562b158" containerName="nova-api-api" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.427714 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c081cab-30c8-4952-8527-65355562b158" containerName="nova-api-log" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.429115 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.431766 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.443320 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.509700 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/009b1923-2cf1-46d6-9cef-7a77c90cdf44-logs\") pod \"nova-api-0\" (UID: \"009b1923-2cf1-46d6-9cef-7a77c90cdf44\") " pod="openstack/nova-api-0" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.510885 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/009b1923-2cf1-46d6-9cef-7a77c90cdf44-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"009b1923-2cf1-46d6-9cef-7a77c90cdf44\") " pod="openstack/nova-api-0" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.510957 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2jqz\" (UniqueName: \"kubernetes.io/projected/009b1923-2cf1-46d6-9cef-7a77c90cdf44-kube-api-access-f2jqz\") pod \"nova-api-0\" (UID: \"009b1923-2cf1-46d6-9cef-7a77c90cdf44\") " pod="openstack/nova-api-0" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.510981 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/009b1923-2cf1-46d6-9cef-7a77c90cdf44-config-data\") pod \"nova-api-0\" (UID: \"009b1923-2cf1-46d6-9cef-7a77c90cdf44\") " pod="openstack/nova-api-0" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.601859 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.612999 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/009b1923-2cf1-46d6-9cef-7a77c90cdf44-logs\") pod \"nova-api-0\" (UID: \"009b1923-2cf1-46d6-9cef-7a77c90cdf44\") " pod="openstack/nova-api-0" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.613081 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/009b1923-2cf1-46d6-9cef-7a77c90cdf44-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"009b1923-2cf1-46d6-9cef-7a77c90cdf44\") " pod="openstack/nova-api-0" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.613119 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2jqz\" (UniqueName: \"kubernetes.io/projected/009b1923-2cf1-46d6-9cef-7a77c90cdf44-kube-api-access-f2jqz\") pod \"nova-api-0\" (UID: \"009b1923-2cf1-46d6-9cef-7a77c90cdf44\") " pod="openstack/nova-api-0" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.613138 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/009b1923-2cf1-46d6-9cef-7a77c90cdf44-config-data\") pod \"nova-api-0\" (UID: \"009b1923-2cf1-46d6-9cef-7a77c90cdf44\") " pod="openstack/nova-api-0" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.616479 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/009b1923-2cf1-46d6-9cef-7a77c90cdf44-config-data\") pod \"nova-api-0\" (UID: \"009b1923-2cf1-46d6-9cef-7a77c90cdf44\") " pod="openstack/nova-api-0" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.616925 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/009b1923-2cf1-46d6-9cef-7a77c90cdf44-logs\") pod \"nova-api-0\" (UID: \"009b1923-2cf1-46d6-9cef-7a77c90cdf44\") " pod="openstack/nova-api-0" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.620098 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/009b1923-2cf1-46d6-9cef-7a77c90cdf44-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"009b1923-2cf1-46d6-9cef-7a77c90cdf44\") " pod="openstack/nova-api-0" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.634861 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2jqz\" (UniqueName: \"kubernetes.io/projected/009b1923-2cf1-46d6-9cef-7a77c90cdf44-kube-api-access-f2jqz\") pod \"nova-api-0\" (UID: \"009b1923-2cf1-46d6-9cef-7a77c90cdf44\") " pod="openstack/nova-api-0" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.752792 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.978502 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4aa280f6-f14f-47de-bbba-8f49229e4ec4","Type":"ContainerStarted","Data":"01261df500f04035637e93506f308f1fcd941eefda1095f190b78293ce07ac1e"} Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.978555 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4aa280f6-f14f-47de-bbba-8f49229e4ec4","Type":"ContainerStarted","Data":"222e6c947b2832af5877014b9ee54911ef4805eeea1f45c1c3f6699210ee4180"} Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.982307 4808 generic.go:334] "Generic (PLEG): container finished" podID="c15a4b91-873b-429f-9b0c-306cda038d5e" containerID="83698ed18252ecc34e23a1609888b85825fbd66f4dc0922676061b11a41342ed" exitCode=0 Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.982370 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323740-2zgrw" event={"ID":"c15a4b91-873b-429f-9b0c-306cda038d5e","Type":"ContainerDied","Data":"83698ed18252ecc34e23a1609888b85825fbd66f4dc0922676061b11a41342ed"} Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.982394 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323740-2zgrw" event={"ID":"c15a4b91-873b-429f-9b0c-306cda038d5e","Type":"ContainerStarted","Data":"76127e7e09cdd463d77f377af929bb65981bc37ffe0b152cfc6ffdb8ed64f39d"} Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.987632 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c66f29c7-9e93-4417-bc3e-b486a0b167cc","Type":"ContainerStarted","Data":"0c3523d23eea77f3a26989699d6139b4952958e86d6780f8cad3f7bce59f1a19"} Oct 02 17:00:01 crc kubenswrapper[4808]: I1002 17:00:01.995553 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.995539682 podStartE2EDuration="1.995539682s" podCreationTimestamp="2025-10-02 17:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 17:00:01.992977193 +0000 UTC m=+1169.318506193" watchObservedRunningTime="2025-10-02 17:00:01.995539682 +0000 UTC m=+1169.321068682" Oct 02 17:00:02 crc kubenswrapper[4808]: I1002 17:00:02.260694 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 17:00:02 crc kubenswrapper[4808]: I1002 17:00:02.996768 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c66f29c7-9e93-4417-bc3e-b486a0b167cc","Type":"ContainerStarted","Data":"c097f25e6215428c150710efb5bf442f8649de49ea19162ee15380e03db6bfbd"} Oct 02 17:00:02 crc kubenswrapper[4808]: I1002 17:00:02.998226 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"009b1923-2cf1-46d6-9cef-7a77c90cdf44","Type":"ContainerStarted","Data":"3bce42cad15e18159a57b6e6e8b3794e0998eaad4b1d358c7ab20ea5fc0fcdcb"} Oct 02 17:00:02 crc kubenswrapper[4808]: I1002 17:00:02.998268 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"009b1923-2cf1-46d6-9cef-7a77c90cdf44","Type":"ContainerStarted","Data":"8f3dfd3e571b272f5d185d494e292385d35fb040845b6682d7ce35a43ec4b10f"} Oct 02 17:00:02 crc kubenswrapper[4808]: I1002 17:00:02.998280 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"009b1923-2cf1-46d6-9cef-7a77c90cdf44","Type":"ContainerStarted","Data":"decc4a5429a4810d4e1885c94bdd542e0ced57b7f0a67a5a2b2904e8a8166c3f"} Oct 02 17:00:03 crc kubenswrapper[4808]: I1002 17:00:03.019347 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.019322261 podStartE2EDuration="2.019322261s" podCreationTimestamp="2025-10-02 17:00:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 17:00:03.013171306 +0000 UTC m=+1170.338700306" watchObservedRunningTime="2025-10-02 17:00:03.019322261 +0000 UTC m=+1170.344851261" Oct 02 17:00:03 crc kubenswrapper[4808]: I1002 17:00:03.409318 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c081cab-30c8-4952-8527-65355562b158" path="/var/lib/kubelet/pods/6c081cab-30c8-4952-8527-65355562b158/volumes" Oct 02 17:00:03 crc kubenswrapper[4808]: I1002 17:00:03.429359 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323740-2zgrw" Oct 02 17:00:03 crc kubenswrapper[4808]: I1002 17:00:03.554884 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c15a4b91-873b-429f-9b0c-306cda038d5e-secret-volume\") pod \"c15a4b91-873b-429f-9b0c-306cda038d5e\" (UID: \"c15a4b91-873b-429f-9b0c-306cda038d5e\") " Oct 02 17:00:03 crc kubenswrapper[4808]: I1002 17:00:03.554943 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c15a4b91-873b-429f-9b0c-306cda038d5e-config-volume\") pod \"c15a4b91-873b-429f-9b0c-306cda038d5e\" (UID: \"c15a4b91-873b-429f-9b0c-306cda038d5e\") " Oct 02 17:00:03 crc kubenswrapper[4808]: I1002 17:00:03.554976 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvpf4\" (UniqueName: \"kubernetes.io/projected/c15a4b91-873b-429f-9b0c-306cda038d5e-kube-api-access-gvpf4\") pod \"c15a4b91-873b-429f-9b0c-306cda038d5e\" (UID: \"c15a4b91-873b-429f-9b0c-306cda038d5e\") " Oct 02 17:00:03 crc kubenswrapper[4808]: I1002 17:00:03.557989 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c15a4b91-873b-429f-9b0c-306cda038d5e-config-volume" (OuterVolumeSpecName: "config-volume") pod "c15a4b91-873b-429f-9b0c-306cda038d5e" (UID: "c15a4b91-873b-429f-9b0c-306cda038d5e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 17:00:03 crc kubenswrapper[4808]: I1002 17:00:03.560965 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c15a4b91-873b-429f-9b0c-306cda038d5e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c15a4b91-873b-429f-9b0c-306cda038d5e" (UID: "c15a4b91-873b-429f-9b0c-306cda038d5e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:03 crc kubenswrapper[4808]: I1002 17:00:03.590431 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c15a4b91-873b-429f-9b0c-306cda038d5e-kube-api-access-gvpf4" (OuterVolumeSpecName: "kube-api-access-gvpf4") pod "c15a4b91-873b-429f-9b0c-306cda038d5e" (UID: "c15a4b91-873b-429f-9b0c-306cda038d5e"). InnerVolumeSpecName "kube-api-access-gvpf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:00:03 crc kubenswrapper[4808]: I1002 17:00:03.656860 4808 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c15a4b91-873b-429f-9b0c-306cda038d5e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:03 crc kubenswrapper[4808]: I1002 17:00:03.656894 4808 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c15a4b91-873b-429f-9b0c-306cda038d5e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:03 crc kubenswrapper[4808]: I1002 17:00:03.656905 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvpf4\" (UniqueName: \"kubernetes.io/projected/c15a4b91-873b-429f-9b0c-306cda038d5e-kube-api-access-gvpf4\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:04 crc kubenswrapper[4808]: I1002 17:00:04.006600 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323740-2zgrw" Oct 02 17:00:04 crc kubenswrapper[4808]: I1002 17:00:04.006628 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323740-2zgrw" event={"ID":"c15a4b91-873b-429f-9b0c-306cda038d5e","Type":"ContainerDied","Data":"76127e7e09cdd463d77f377af929bb65981bc37ffe0b152cfc6ffdb8ed64f39d"} Oct 02 17:00:04 crc kubenswrapper[4808]: I1002 17:00:04.006665 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76127e7e09cdd463d77f377af929bb65981bc37ffe0b152cfc6ffdb8ed64f39d" Oct 02 17:00:05 crc kubenswrapper[4808]: I1002 17:00:05.633963 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 17:00:06 crc kubenswrapper[4808]: I1002 17:00:06.040377 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c66f29c7-9e93-4417-bc3e-b486a0b167cc","Type":"ContainerStarted","Data":"95e245e2a282a68d227f87f0a4202fabdd57814fc3e943d366c73a22d098be10"} Oct 02 17:00:06 crc kubenswrapper[4808]: I1002 17:00:06.041668 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 17:00:06 crc kubenswrapper[4808]: I1002 17:00:06.075606 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.041412632 podStartE2EDuration="8.075582914s" podCreationTimestamp="2025-10-02 16:59:58 +0000 UTC" firstStartedPulling="2025-10-02 16:59:59.808962114 +0000 UTC m=+1167.134491114" lastFinishedPulling="2025-10-02 17:00:04.843132386 +0000 UTC m=+1172.168661396" observedRunningTime="2025-10-02 17:00:06.071510565 +0000 UTC m=+1173.397039605" watchObservedRunningTime="2025-10-02 17:00:06.075582914 +0000 UTC m=+1173.401111944" Oct 02 17:00:06 crc kubenswrapper[4808]: I1002 17:00:06.299213 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 17:00:06 crc kubenswrapper[4808]: I1002 17:00:06.299770 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 17:00:06 crc kubenswrapper[4808]: I1002 17:00:06.309885 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 02 17:00:07 crc kubenswrapper[4808]: I1002 17:00:07.310578 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8d7a4dd0-7af5-4d01-b888-14ce54434b2d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.178:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 17:00:07 crc kubenswrapper[4808]: I1002 17:00:07.310645 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8d7a4dd0-7af5-4d01-b888-14ce54434b2d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.178:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 17:00:10 crc kubenswrapper[4808]: I1002 17:00:10.633867 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 17:00:10 crc kubenswrapper[4808]: I1002 17:00:10.682989 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 17:00:10 crc kubenswrapper[4808]: I1002 17:00:10.707586 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 17:00:11 crc kubenswrapper[4808]: I1002 17:00:11.114795 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 17:00:11 crc kubenswrapper[4808]: I1002 17:00:11.753449 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 17:00:11 crc kubenswrapper[4808]: I1002 17:00:11.753848 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 17:00:12 crc kubenswrapper[4808]: I1002 17:00:12.836400 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="009b1923-2cf1-46d6-9cef-7a77c90cdf44" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.182:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 17:00:12 crc kubenswrapper[4808]: I1002 17:00:12.836420 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="009b1923-2cf1-46d6-9cef-7a77c90cdf44" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.182:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 17:00:16 crc kubenswrapper[4808]: I1002 17:00:16.308337 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 17:00:16 crc kubenswrapper[4808]: I1002 17:00:16.309165 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 17:00:16 crc kubenswrapper[4808]: I1002 17:00:16.317695 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 17:00:16 crc kubenswrapper[4808]: I1002 17:00:16.318522 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.139580 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.167250 4808 generic.go:334] "Generic (PLEG): container finished" podID="7cd4a7e0-6aa4-4418-889a-2112878249bd" containerID="89aeb4bdb27b5d5cc32c7d7802c649112b39faa31e0ffae6369762cd3ac6f778" exitCode=137 Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.167297 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7cd4a7e0-6aa4-4418-889a-2112878249bd","Type":"ContainerDied","Data":"89aeb4bdb27b5d5cc32c7d7802c649112b39faa31e0ffae6369762cd3ac6f778"} Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.167325 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7cd4a7e0-6aa4-4418-889a-2112878249bd","Type":"ContainerDied","Data":"3cfe8b5418e423e672ce59e3e16d26333e15bbbe203c3ed3a46d693aca1e3225"} Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.167345 4808 scope.go:117] "RemoveContainer" containerID="89aeb4bdb27b5d5cc32c7d7802c649112b39faa31e0ffae6369762cd3ac6f778" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.167491 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.203346 4808 scope.go:117] "RemoveContainer" containerID="89aeb4bdb27b5d5cc32c7d7802c649112b39faa31e0ffae6369762cd3ac6f778" Oct 02 17:00:19 crc kubenswrapper[4808]: E1002 17:00:19.204920 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89aeb4bdb27b5d5cc32c7d7802c649112b39faa31e0ffae6369762cd3ac6f778\": container with ID starting with 89aeb4bdb27b5d5cc32c7d7802c649112b39faa31e0ffae6369762cd3ac6f778 not found: ID does not exist" containerID="89aeb4bdb27b5d5cc32c7d7802c649112b39faa31e0ffae6369762cd3ac6f778" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.204968 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89aeb4bdb27b5d5cc32c7d7802c649112b39faa31e0ffae6369762cd3ac6f778"} err="failed to get container status \"89aeb4bdb27b5d5cc32c7d7802c649112b39faa31e0ffae6369762cd3ac6f778\": rpc error: code = NotFound desc = could not find container \"89aeb4bdb27b5d5cc32c7d7802c649112b39faa31e0ffae6369762cd3ac6f778\": container with ID starting with 89aeb4bdb27b5d5cc32c7d7802c649112b39faa31e0ffae6369762cd3ac6f778 not found: ID does not exist" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.274493 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cd4a7e0-6aa4-4418-889a-2112878249bd-combined-ca-bundle\") pod \"7cd4a7e0-6aa4-4418-889a-2112878249bd\" (UID: \"7cd4a7e0-6aa4-4418-889a-2112878249bd\") " Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.275204 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2jb2\" (UniqueName: \"kubernetes.io/projected/7cd4a7e0-6aa4-4418-889a-2112878249bd-kube-api-access-l2jb2\") pod \"7cd4a7e0-6aa4-4418-889a-2112878249bd\" (UID: \"7cd4a7e0-6aa4-4418-889a-2112878249bd\") " Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.275627 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cd4a7e0-6aa4-4418-889a-2112878249bd-config-data\") pod \"7cd4a7e0-6aa4-4418-889a-2112878249bd\" (UID: \"7cd4a7e0-6aa4-4418-889a-2112878249bd\") " Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.280929 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cd4a7e0-6aa4-4418-889a-2112878249bd-kube-api-access-l2jb2" (OuterVolumeSpecName: "kube-api-access-l2jb2") pod "7cd4a7e0-6aa4-4418-889a-2112878249bd" (UID: "7cd4a7e0-6aa4-4418-889a-2112878249bd"). InnerVolumeSpecName "kube-api-access-l2jb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.302716 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cd4a7e0-6aa4-4418-889a-2112878249bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7cd4a7e0-6aa4-4418-889a-2112878249bd" (UID: "7cd4a7e0-6aa4-4418-889a-2112878249bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.313503 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cd4a7e0-6aa4-4418-889a-2112878249bd-config-data" (OuterVolumeSpecName: "config-data") pod "7cd4a7e0-6aa4-4418-889a-2112878249bd" (UID: "7cd4a7e0-6aa4-4418-889a-2112878249bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.378779 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cd4a7e0-6aa4-4418-889a-2112878249bd-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.378819 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cd4a7e0-6aa4-4418-889a-2112878249bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.378835 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2jb2\" (UniqueName: \"kubernetes.io/projected/7cd4a7e0-6aa4-4418-889a-2112878249bd-kube-api-access-l2jb2\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.489597 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.497165 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.516927 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 17:00:19 crc kubenswrapper[4808]: E1002 17:00:19.517365 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cd4a7e0-6aa4-4418-889a-2112878249bd" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.517387 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cd4a7e0-6aa4-4418-889a-2112878249bd" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 17:00:19 crc kubenswrapper[4808]: E1002 17:00:19.517408 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c15a4b91-873b-429f-9b0c-306cda038d5e" containerName="collect-profiles" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.517416 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c15a4b91-873b-429f-9b0c-306cda038d5e" containerName="collect-profiles" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.517623 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c15a4b91-873b-429f-9b0c-306cda038d5e" containerName="collect-profiles" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.517650 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cd4a7e0-6aa4-4418-889a-2112878249bd" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.518449 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.521100 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.521140 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.521625 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.541116 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.582995 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d89cc77e-ed29-4940-8db0-b7c32499d59e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d89cc77e-ed29-4940-8db0-b7c32499d59e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.583052 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d89cc77e-ed29-4940-8db0-b7c32499d59e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d89cc77e-ed29-4940-8db0-b7c32499d59e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.583163 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d89cc77e-ed29-4940-8db0-b7c32499d59e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d89cc77e-ed29-4940-8db0-b7c32499d59e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.583220 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d89cc77e-ed29-4940-8db0-b7c32499d59e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d89cc77e-ed29-4940-8db0-b7c32499d59e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.583274 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2l56\" (UniqueName: \"kubernetes.io/projected/d89cc77e-ed29-4940-8db0-b7c32499d59e-kube-api-access-f2l56\") pod \"nova-cell1-novncproxy-0\" (UID: \"d89cc77e-ed29-4940-8db0-b7c32499d59e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.685067 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d89cc77e-ed29-4940-8db0-b7c32499d59e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d89cc77e-ed29-4940-8db0-b7c32499d59e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.685128 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d89cc77e-ed29-4940-8db0-b7c32499d59e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d89cc77e-ed29-4940-8db0-b7c32499d59e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.685162 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2l56\" (UniqueName: \"kubernetes.io/projected/d89cc77e-ed29-4940-8db0-b7c32499d59e-kube-api-access-f2l56\") pod \"nova-cell1-novncproxy-0\" (UID: \"d89cc77e-ed29-4940-8db0-b7c32499d59e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.685277 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d89cc77e-ed29-4940-8db0-b7c32499d59e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d89cc77e-ed29-4940-8db0-b7c32499d59e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.685482 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d89cc77e-ed29-4940-8db0-b7c32499d59e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d89cc77e-ed29-4940-8db0-b7c32499d59e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.688888 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d89cc77e-ed29-4940-8db0-b7c32499d59e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d89cc77e-ed29-4940-8db0-b7c32499d59e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.689184 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d89cc77e-ed29-4940-8db0-b7c32499d59e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d89cc77e-ed29-4940-8db0-b7c32499d59e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.689823 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d89cc77e-ed29-4940-8db0-b7c32499d59e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d89cc77e-ed29-4940-8db0-b7c32499d59e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.690982 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d89cc77e-ed29-4940-8db0-b7c32499d59e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d89cc77e-ed29-4940-8db0-b7c32499d59e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.715700 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2l56\" (UniqueName: \"kubernetes.io/projected/d89cc77e-ed29-4940-8db0-b7c32499d59e-kube-api-access-f2l56\") pod \"nova-cell1-novncproxy-0\" (UID: \"d89cc77e-ed29-4940-8db0-b7c32499d59e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:19 crc kubenswrapper[4808]: I1002 17:00:19.842037 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:20 crc kubenswrapper[4808]: I1002 17:00:20.158939 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 17:00:20 crc kubenswrapper[4808]: I1002 17:00:20.176796 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d89cc77e-ed29-4940-8db0-b7c32499d59e","Type":"ContainerStarted","Data":"c704e29b9e3a9504a71b2922599f70654daf3c9a9cbc52239f9b9389cdb31b5f"} Oct 02 17:00:21 crc kubenswrapper[4808]: I1002 17:00:21.200525 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d89cc77e-ed29-4940-8db0-b7c32499d59e","Type":"ContainerStarted","Data":"3debcd78a09e8410fa2c60afe0e2df66a849a8fc13d8b692a668835c11d7330b"} Oct 02 17:00:21 crc kubenswrapper[4808]: I1002 17:00:21.241397 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.241335162 podStartE2EDuration="2.241335162s" podCreationTimestamp="2025-10-02 17:00:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 17:00:21.240264853 +0000 UTC m=+1188.565793883" watchObservedRunningTime="2025-10-02 17:00:21.241335162 +0000 UTC m=+1188.566864192" Oct 02 17:00:21 crc kubenswrapper[4808]: I1002 17:00:21.414428 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cd4a7e0-6aa4-4418-889a-2112878249bd" path="/var/lib/kubelet/pods/7cd4a7e0-6aa4-4418-889a-2112878249bd/volumes" Oct 02 17:00:21 crc kubenswrapper[4808]: I1002 17:00:21.758797 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 17:00:21 crc kubenswrapper[4808]: I1002 17:00:21.758898 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 17:00:21 crc kubenswrapper[4808]: I1002 17:00:21.759565 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 17:00:21 crc kubenswrapper[4808]: I1002 17:00:21.759634 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 17:00:21 crc kubenswrapper[4808]: I1002 17:00:21.763542 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 17:00:21 crc kubenswrapper[4808]: I1002 17:00:21.766132 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 17:00:21 crc kubenswrapper[4808]: I1002 17:00:21.983585 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-7rnjh"] Oct 02 17:00:21 crc kubenswrapper[4808]: I1002 17:00:21.987286 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:00:22 crc kubenswrapper[4808]: I1002 17:00:22.004490 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-7rnjh"] Oct 02 17:00:22 crc kubenswrapper[4808]: I1002 17:00:22.132075 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6snvb\" (UniqueName: \"kubernetes.io/projected/e11ba3da-0594-456b-8cb1-2837d0422c2b-kube-api-access-6snvb\") pod \"dnsmasq-dns-5b856c5697-7rnjh\" (UID: \"e11ba3da-0594-456b-8cb1-2837d0422c2b\") " pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:00:22 crc kubenswrapper[4808]: I1002 17:00:22.132351 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-ovsdbserver-nb\") pod \"dnsmasq-dns-5b856c5697-7rnjh\" (UID: \"e11ba3da-0594-456b-8cb1-2837d0422c2b\") " pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:00:22 crc kubenswrapper[4808]: I1002 17:00:22.132466 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-ovsdbserver-sb\") pod \"dnsmasq-dns-5b856c5697-7rnjh\" (UID: \"e11ba3da-0594-456b-8cb1-2837d0422c2b\") " pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:00:22 crc kubenswrapper[4808]: I1002 17:00:22.132592 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-dns-svc\") pod \"dnsmasq-dns-5b856c5697-7rnjh\" (UID: \"e11ba3da-0594-456b-8cb1-2837d0422c2b\") " pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:00:22 crc kubenswrapper[4808]: I1002 17:00:22.132663 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-config\") pod \"dnsmasq-dns-5b856c5697-7rnjh\" (UID: \"e11ba3da-0594-456b-8cb1-2837d0422c2b\") " pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:00:22 crc kubenswrapper[4808]: I1002 17:00:22.234571 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6snvb\" (UniqueName: \"kubernetes.io/projected/e11ba3da-0594-456b-8cb1-2837d0422c2b-kube-api-access-6snvb\") pod \"dnsmasq-dns-5b856c5697-7rnjh\" (UID: \"e11ba3da-0594-456b-8cb1-2837d0422c2b\") " pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:00:22 crc kubenswrapper[4808]: I1002 17:00:22.234943 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-ovsdbserver-nb\") pod \"dnsmasq-dns-5b856c5697-7rnjh\" (UID: \"e11ba3da-0594-456b-8cb1-2837d0422c2b\") " pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:00:22 crc kubenswrapper[4808]: I1002 17:00:22.235001 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-ovsdbserver-sb\") pod \"dnsmasq-dns-5b856c5697-7rnjh\" (UID: \"e11ba3da-0594-456b-8cb1-2837d0422c2b\") " pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:00:22 crc kubenswrapper[4808]: I1002 17:00:22.235077 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-dns-svc\") pod \"dnsmasq-dns-5b856c5697-7rnjh\" (UID: \"e11ba3da-0594-456b-8cb1-2837d0422c2b\") " pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:00:22 crc kubenswrapper[4808]: I1002 17:00:22.235106 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-config\") pod \"dnsmasq-dns-5b856c5697-7rnjh\" (UID: \"e11ba3da-0594-456b-8cb1-2837d0422c2b\") " pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:00:22 crc kubenswrapper[4808]: I1002 17:00:22.235860 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-config\") pod \"dnsmasq-dns-5b856c5697-7rnjh\" (UID: \"e11ba3da-0594-456b-8cb1-2837d0422c2b\") " pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:00:22 crc kubenswrapper[4808]: I1002 17:00:22.237310 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-ovsdbserver-nb\") pod \"dnsmasq-dns-5b856c5697-7rnjh\" (UID: \"e11ba3da-0594-456b-8cb1-2837d0422c2b\") " pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:00:22 crc kubenswrapper[4808]: I1002 17:00:22.237819 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-ovsdbserver-sb\") pod \"dnsmasq-dns-5b856c5697-7rnjh\" (UID: \"e11ba3da-0594-456b-8cb1-2837d0422c2b\") " pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:00:22 crc kubenswrapper[4808]: I1002 17:00:22.238318 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-dns-svc\") pod \"dnsmasq-dns-5b856c5697-7rnjh\" (UID: \"e11ba3da-0594-456b-8cb1-2837d0422c2b\") " pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:00:22 crc kubenswrapper[4808]: I1002 17:00:22.261032 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6snvb\" (UniqueName: \"kubernetes.io/projected/e11ba3da-0594-456b-8cb1-2837d0422c2b-kube-api-access-6snvb\") pod \"dnsmasq-dns-5b856c5697-7rnjh\" (UID: \"e11ba3da-0594-456b-8cb1-2837d0422c2b\") " pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:00:22 crc kubenswrapper[4808]: I1002 17:00:22.325979 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:00:22 crc kubenswrapper[4808]: I1002 17:00:22.812262 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-7rnjh"] Oct 02 17:00:22 crc kubenswrapper[4808]: W1002 17:00:22.818323 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode11ba3da_0594_456b_8cb1_2837d0422c2b.slice/crio-ec3bef622f9667e59968ba9631c6c7c47607292849c8671b6bc952d00bad8cbc WatchSource:0}: Error finding container ec3bef622f9667e59968ba9631c6c7c47607292849c8671b6bc952d00bad8cbc: Status 404 returned error can't find the container with id ec3bef622f9667e59968ba9631c6c7c47607292849c8671b6bc952d00bad8cbc Oct 02 17:00:23 crc kubenswrapper[4808]: I1002 17:00:23.224749 4808 generic.go:334] "Generic (PLEG): container finished" podID="e11ba3da-0594-456b-8cb1-2837d0422c2b" containerID="665dd85ef26f170f6e0fc1a9cc8d4962dda354e1245fdf60001f55028451190e" exitCode=0 Oct 02 17:00:23 crc kubenswrapper[4808]: I1002 17:00:23.226399 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" event={"ID":"e11ba3da-0594-456b-8cb1-2837d0422c2b","Type":"ContainerDied","Data":"665dd85ef26f170f6e0fc1a9cc8d4962dda354e1245fdf60001f55028451190e"} Oct 02 17:00:23 crc kubenswrapper[4808]: I1002 17:00:23.226467 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" event={"ID":"e11ba3da-0594-456b-8cb1-2837d0422c2b","Type":"ContainerStarted","Data":"ec3bef622f9667e59968ba9631c6c7c47607292849c8671b6bc952d00bad8cbc"} Oct 02 17:00:24 crc kubenswrapper[4808]: I1002 17:00:24.215402 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 17:00:24 crc kubenswrapper[4808]: I1002 17:00:24.239123 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" event={"ID":"e11ba3da-0594-456b-8cb1-2837d0422c2b","Type":"ContainerStarted","Data":"61c2aadd62ddfbaaf03579537207f98e73b0f61d970ecc126ce0fb618e98079f"} Oct 02 17:00:24 crc kubenswrapper[4808]: I1002 17:00:24.239453 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="009b1923-2cf1-46d6-9cef-7a77c90cdf44" containerName="nova-api-log" containerID="cri-o://8f3dfd3e571b272f5d185d494e292385d35fb040845b6682d7ce35a43ec4b10f" gracePeriod=30 Oct 02 17:00:24 crc kubenswrapper[4808]: I1002 17:00:24.239524 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="009b1923-2cf1-46d6-9cef-7a77c90cdf44" containerName="nova-api-api" containerID="cri-o://3bce42cad15e18159a57b6e6e8b3794e0998eaad4b1d358c7ab20ea5fc0fcdcb" gracePeriod=30 Oct 02 17:00:24 crc kubenswrapper[4808]: I1002 17:00:24.264961 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" podStartSLOduration=3.26494011 podStartE2EDuration="3.26494011s" podCreationTimestamp="2025-10-02 17:00:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 17:00:24.264671093 +0000 UTC m=+1191.590200093" watchObservedRunningTime="2025-10-02 17:00:24.26494011 +0000 UTC m=+1191.590469110" Oct 02 17:00:24 crc kubenswrapper[4808]: I1002 17:00:24.843214 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:25 crc kubenswrapper[4808]: I1002 17:00:25.251438 4808 generic.go:334] "Generic (PLEG): container finished" podID="009b1923-2cf1-46d6-9cef-7a77c90cdf44" containerID="8f3dfd3e571b272f5d185d494e292385d35fb040845b6682d7ce35a43ec4b10f" exitCode=143 Oct 02 17:00:25 crc kubenswrapper[4808]: I1002 17:00:25.251492 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"009b1923-2cf1-46d6-9cef-7a77c90cdf44","Type":"ContainerDied","Data":"8f3dfd3e571b272f5d185d494e292385d35fb040845b6682d7ce35a43ec4b10f"} Oct 02 17:00:25 crc kubenswrapper[4808]: I1002 17:00:25.251776 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:00:27 crc kubenswrapper[4808]: I1002 17:00:27.885574 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.048589 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/009b1923-2cf1-46d6-9cef-7a77c90cdf44-combined-ca-bundle\") pod \"009b1923-2cf1-46d6-9cef-7a77c90cdf44\" (UID: \"009b1923-2cf1-46d6-9cef-7a77c90cdf44\") " Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.049221 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/009b1923-2cf1-46d6-9cef-7a77c90cdf44-config-data\") pod \"009b1923-2cf1-46d6-9cef-7a77c90cdf44\" (UID: \"009b1923-2cf1-46d6-9cef-7a77c90cdf44\") " Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.049351 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2jqz\" (UniqueName: \"kubernetes.io/projected/009b1923-2cf1-46d6-9cef-7a77c90cdf44-kube-api-access-f2jqz\") pod \"009b1923-2cf1-46d6-9cef-7a77c90cdf44\" (UID: \"009b1923-2cf1-46d6-9cef-7a77c90cdf44\") " Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.049407 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/009b1923-2cf1-46d6-9cef-7a77c90cdf44-logs\") pod \"009b1923-2cf1-46d6-9cef-7a77c90cdf44\" (UID: \"009b1923-2cf1-46d6-9cef-7a77c90cdf44\") " Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.052411 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/009b1923-2cf1-46d6-9cef-7a77c90cdf44-logs" (OuterVolumeSpecName: "logs") pod "009b1923-2cf1-46d6-9cef-7a77c90cdf44" (UID: "009b1923-2cf1-46d6-9cef-7a77c90cdf44"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.059933 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/009b1923-2cf1-46d6-9cef-7a77c90cdf44-kube-api-access-f2jqz" (OuterVolumeSpecName: "kube-api-access-f2jqz") pod "009b1923-2cf1-46d6-9cef-7a77c90cdf44" (UID: "009b1923-2cf1-46d6-9cef-7a77c90cdf44"). InnerVolumeSpecName "kube-api-access-f2jqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.084513 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/009b1923-2cf1-46d6-9cef-7a77c90cdf44-config-data" (OuterVolumeSpecName: "config-data") pod "009b1923-2cf1-46d6-9cef-7a77c90cdf44" (UID: "009b1923-2cf1-46d6-9cef-7a77c90cdf44"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.116803 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/009b1923-2cf1-46d6-9cef-7a77c90cdf44-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "009b1923-2cf1-46d6-9cef-7a77c90cdf44" (UID: "009b1923-2cf1-46d6-9cef-7a77c90cdf44"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.152274 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/009b1923-2cf1-46d6-9cef-7a77c90cdf44-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.152317 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2jqz\" (UniqueName: \"kubernetes.io/projected/009b1923-2cf1-46d6-9cef-7a77c90cdf44-kube-api-access-f2jqz\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.152328 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/009b1923-2cf1-46d6-9cef-7a77c90cdf44-logs\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.152336 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/009b1923-2cf1-46d6-9cef-7a77c90cdf44-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.286052 4808 generic.go:334] "Generic (PLEG): container finished" podID="009b1923-2cf1-46d6-9cef-7a77c90cdf44" containerID="3bce42cad15e18159a57b6e6e8b3794e0998eaad4b1d358c7ab20ea5fc0fcdcb" exitCode=0 Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.286092 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"009b1923-2cf1-46d6-9cef-7a77c90cdf44","Type":"ContainerDied","Data":"3bce42cad15e18159a57b6e6e8b3794e0998eaad4b1d358c7ab20ea5fc0fcdcb"} Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.286117 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"009b1923-2cf1-46d6-9cef-7a77c90cdf44","Type":"ContainerDied","Data":"decc4a5429a4810d4e1885c94bdd542e0ced57b7f0a67a5a2b2904e8a8166c3f"} Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.286141 4808 scope.go:117] "RemoveContainer" containerID="3bce42cad15e18159a57b6e6e8b3794e0998eaad4b1d358c7ab20ea5fc0fcdcb" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.286159 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.313198 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.320844 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.332117 4808 scope.go:117] "RemoveContainer" containerID="8f3dfd3e571b272f5d185d494e292385d35fb040845b6682d7ce35a43ec4b10f" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.335284 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 17:00:28 crc kubenswrapper[4808]: E1002 17:00:28.335674 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="009b1923-2cf1-46d6-9cef-7a77c90cdf44" containerName="nova-api-log" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.335690 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="009b1923-2cf1-46d6-9cef-7a77c90cdf44" containerName="nova-api-log" Oct 02 17:00:28 crc kubenswrapper[4808]: E1002 17:00:28.335720 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="009b1923-2cf1-46d6-9cef-7a77c90cdf44" containerName="nova-api-api" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.335727 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="009b1923-2cf1-46d6-9cef-7a77c90cdf44" containerName="nova-api-api" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.335880 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="009b1923-2cf1-46d6-9cef-7a77c90cdf44" containerName="nova-api-log" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.335900 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="009b1923-2cf1-46d6-9cef-7a77c90cdf44" containerName="nova-api-api" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.336825 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.339531 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.339689 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.339734 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.344902 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.353995 4808 scope.go:117] "RemoveContainer" containerID="3bce42cad15e18159a57b6e6e8b3794e0998eaad4b1d358c7ab20ea5fc0fcdcb" Oct 02 17:00:28 crc kubenswrapper[4808]: E1002 17:00:28.354432 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bce42cad15e18159a57b6e6e8b3794e0998eaad4b1d358c7ab20ea5fc0fcdcb\": container with ID starting with 3bce42cad15e18159a57b6e6e8b3794e0998eaad4b1d358c7ab20ea5fc0fcdcb not found: ID does not exist" containerID="3bce42cad15e18159a57b6e6e8b3794e0998eaad4b1d358c7ab20ea5fc0fcdcb" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.354472 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bce42cad15e18159a57b6e6e8b3794e0998eaad4b1d358c7ab20ea5fc0fcdcb"} err="failed to get container status \"3bce42cad15e18159a57b6e6e8b3794e0998eaad4b1d358c7ab20ea5fc0fcdcb\": rpc error: code = NotFound desc = could not find container \"3bce42cad15e18159a57b6e6e8b3794e0998eaad4b1d358c7ab20ea5fc0fcdcb\": container with ID starting with 3bce42cad15e18159a57b6e6e8b3794e0998eaad4b1d358c7ab20ea5fc0fcdcb not found: ID does not exist" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.354504 4808 scope.go:117] "RemoveContainer" containerID="8f3dfd3e571b272f5d185d494e292385d35fb040845b6682d7ce35a43ec4b10f" Oct 02 17:00:28 crc kubenswrapper[4808]: E1002 17:00:28.354817 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f3dfd3e571b272f5d185d494e292385d35fb040845b6682d7ce35a43ec4b10f\": container with ID starting with 8f3dfd3e571b272f5d185d494e292385d35fb040845b6682d7ce35a43ec4b10f not found: ID does not exist" containerID="8f3dfd3e571b272f5d185d494e292385d35fb040845b6682d7ce35a43ec4b10f" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.354851 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f3dfd3e571b272f5d185d494e292385d35fb040845b6682d7ce35a43ec4b10f"} err="failed to get container status \"8f3dfd3e571b272f5d185d494e292385d35fb040845b6682d7ce35a43ec4b10f\": rpc error: code = NotFound desc = could not find container \"8f3dfd3e571b272f5d185d494e292385d35fb040845b6682d7ce35a43ec4b10f\": container with ID starting with 8f3dfd3e571b272f5d185d494e292385d35fb040845b6682d7ce35a43ec4b10f not found: ID does not exist" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.456903 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-public-tls-certs\") pod \"nova-api-0\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.457024 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-config-data\") pod \"nova-api-0\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.457049 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.457083 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.457099 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd9jg\" (UniqueName: \"kubernetes.io/projected/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-kube-api-access-qd9jg\") pod \"nova-api-0\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.457135 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-logs\") pod \"nova-api-0\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.559647 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.559725 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd9jg\" (UniqueName: \"kubernetes.io/projected/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-kube-api-access-qd9jg\") pod \"nova-api-0\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.559807 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-logs\") pod \"nova-api-0\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.559862 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-public-tls-certs\") pod \"nova-api-0\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.560573 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-logs\") pod \"nova-api-0\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.560844 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-config-data\") pod \"nova-api-0\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.560896 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.564495 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.566932 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.568461 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-config-data\") pod \"nova-api-0\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.568868 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-public-tls-certs\") pod \"nova-api-0\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.587872 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd9jg\" (UniqueName: \"kubernetes.io/projected/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-kube-api-access-qd9jg\") pod \"nova-api-0\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.668165 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 17:00:28 crc kubenswrapper[4808]: I1002 17:00:28.957281 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 17:00:29 crc kubenswrapper[4808]: I1002 17:00:29.305667 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c8a3b077-ac37-4c23-ad69-2916a00ce4b5","Type":"ContainerStarted","Data":"c3af4d6aaf63056ce1e2dd7d19db4405a1c4ff06857cb2311174c96ce9129041"} Oct 02 17:00:29 crc kubenswrapper[4808]: I1002 17:00:29.306011 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c8a3b077-ac37-4c23-ad69-2916a00ce4b5","Type":"ContainerStarted","Data":"201a7428185006e7ab205cff2f9cb249d373ee597bd924728bcef649857974bd"} Oct 02 17:00:29 crc kubenswrapper[4808]: I1002 17:00:29.345328 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 17:00:29 crc kubenswrapper[4808]: I1002 17:00:29.420823 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="009b1923-2cf1-46d6-9cef-7a77c90cdf44" path="/var/lib/kubelet/pods/009b1923-2cf1-46d6-9cef-7a77c90cdf44/volumes" Oct 02 17:00:29 crc kubenswrapper[4808]: I1002 17:00:29.843402 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:29 crc kubenswrapper[4808]: I1002 17:00:29.873178 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.319249 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c8a3b077-ac37-4c23-ad69-2916a00ce4b5","Type":"ContainerStarted","Data":"557c30fafca450d2bb3f4df11db3e147992fe0f858ac738021c96fe43f85feb2"} Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.339125 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.339104128 podStartE2EDuration="2.339104128s" podCreationTimestamp="2025-10-02 17:00:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 17:00:30.337546937 +0000 UTC m=+1197.663075977" watchObservedRunningTime="2025-10-02 17:00:30.339104128 +0000 UTC m=+1197.664633158" Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.352292 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.553641 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-jkwct"] Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.554675 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-jkwct" Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.557438 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.557750 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.575795 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-jkwct"] Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.617966 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-scripts\") pod \"nova-cell1-cell-mapping-jkwct\" (UID: \"802da733-8c7c-450b-a9c6-7fda3f2d2fbc\") " pod="openstack/nova-cell1-cell-mapping-jkwct" Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.618003 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-config-data\") pod \"nova-cell1-cell-mapping-jkwct\" (UID: \"802da733-8c7c-450b-a9c6-7fda3f2d2fbc\") " pod="openstack/nova-cell1-cell-mapping-jkwct" Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.618064 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpt6n\" (UniqueName: \"kubernetes.io/projected/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-kube-api-access-zpt6n\") pod \"nova-cell1-cell-mapping-jkwct\" (UID: \"802da733-8c7c-450b-a9c6-7fda3f2d2fbc\") " pod="openstack/nova-cell1-cell-mapping-jkwct" Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.618115 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-jkwct\" (UID: \"802da733-8c7c-450b-a9c6-7fda3f2d2fbc\") " pod="openstack/nova-cell1-cell-mapping-jkwct" Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.719524 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-scripts\") pod \"nova-cell1-cell-mapping-jkwct\" (UID: \"802da733-8c7c-450b-a9c6-7fda3f2d2fbc\") " pod="openstack/nova-cell1-cell-mapping-jkwct" Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.719573 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-config-data\") pod \"nova-cell1-cell-mapping-jkwct\" (UID: \"802da733-8c7c-450b-a9c6-7fda3f2d2fbc\") " pod="openstack/nova-cell1-cell-mapping-jkwct" Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.719634 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpt6n\" (UniqueName: \"kubernetes.io/projected/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-kube-api-access-zpt6n\") pod \"nova-cell1-cell-mapping-jkwct\" (UID: \"802da733-8c7c-450b-a9c6-7fda3f2d2fbc\") " pod="openstack/nova-cell1-cell-mapping-jkwct" Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.719691 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-jkwct\" (UID: \"802da733-8c7c-450b-a9c6-7fda3f2d2fbc\") " pod="openstack/nova-cell1-cell-mapping-jkwct" Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.725487 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-jkwct\" (UID: \"802da733-8c7c-450b-a9c6-7fda3f2d2fbc\") " pod="openstack/nova-cell1-cell-mapping-jkwct" Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.725698 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-config-data\") pod \"nova-cell1-cell-mapping-jkwct\" (UID: \"802da733-8c7c-450b-a9c6-7fda3f2d2fbc\") " pod="openstack/nova-cell1-cell-mapping-jkwct" Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.726100 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-scripts\") pod \"nova-cell1-cell-mapping-jkwct\" (UID: \"802da733-8c7c-450b-a9c6-7fda3f2d2fbc\") " pod="openstack/nova-cell1-cell-mapping-jkwct" Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.758812 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpt6n\" (UniqueName: \"kubernetes.io/projected/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-kube-api-access-zpt6n\") pod \"nova-cell1-cell-mapping-jkwct\" (UID: \"802da733-8c7c-450b-a9c6-7fda3f2d2fbc\") " pod="openstack/nova-cell1-cell-mapping-jkwct" Oct 02 17:00:30 crc kubenswrapper[4808]: I1002 17:00:30.871546 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-jkwct" Oct 02 17:00:31 crc kubenswrapper[4808]: I1002 17:00:31.338500 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-jkwct"] Oct 02 17:00:31 crc kubenswrapper[4808]: W1002 17:00:31.351649 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod802da733_8c7c_450b_a9c6_7fda3f2d2fbc.slice/crio-cb9778d80c516318bf9db29e486337092274f03159b04c42de8b98505bed1e98 WatchSource:0}: Error finding container cb9778d80c516318bf9db29e486337092274f03159b04c42de8b98505bed1e98: Status 404 returned error can't find the container with id cb9778d80c516318bf9db29e486337092274f03159b04c42de8b98505bed1e98 Oct 02 17:00:32 crc kubenswrapper[4808]: I1002 17:00:32.327432 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:00:32 crc kubenswrapper[4808]: I1002 17:00:32.345484 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-jkwct" event={"ID":"802da733-8c7c-450b-a9c6-7fda3f2d2fbc","Type":"ContainerStarted","Data":"9a65749ab5db1a8196465d4bae65ab12cec790701140fd3591a447236d973492"} Oct 02 17:00:32 crc kubenswrapper[4808]: I1002 17:00:32.345521 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-jkwct" event={"ID":"802da733-8c7c-450b-a9c6-7fda3f2d2fbc","Type":"ContainerStarted","Data":"cb9778d80c516318bf9db29e486337092274f03159b04c42de8b98505bed1e98"} Oct 02 17:00:32 crc kubenswrapper[4808]: I1002 17:00:32.372685 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-jkwct" podStartSLOduration=2.372670172 podStartE2EDuration="2.372670172s" podCreationTimestamp="2025-10-02 17:00:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 17:00:32.362646124 +0000 UTC m=+1199.688175134" watchObservedRunningTime="2025-10-02 17:00:32.372670172 +0000 UTC m=+1199.698199172" Oct 02 17:00:32 crc kubenswrapper[4808]: I1002 17:00:32.408913 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-85kbf"] Oct 02 17:00:32 crc kubenswrapper[4808]: I1002 17:00:32.410873 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-566b5b7845-85kbf" podUID="fe4d6d85-749d-4ca8-a962-e49678672538" containerName="dnsmasq-dns" containerID="cri-o://14265a8079d6478a47893642811ce11637483a21a6d50eb089c39f67ea514e48" gracePeriod=10 Oct 02 17:00:32 crc kubenswrapper[4808]: I1002 17:00:32.908088 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.074532 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-ovsdbserver-sb\") pod \"fe4d6d85-749d-4ca8-a962-e49678672538\" (UID: \"fe4d6d85-749d-4ca8-a962-e49678672538\") " Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.075184 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-dns-svc\") pod \"fe4d6d85-749d-4ca8-a962-e49678672538\" (UID: \"fe4d6d85-749d-4ca8-a962-e49678672538\") " Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.075302 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-config\") pod \"fe4d6d85-749d-4ca8-a962-e49678672538\" (UID: \"fe4d6d85-749d-4ca8-a962-e49678672538\") " Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.075715 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dmm8\" (UniqueName: \"kubernetes.io/projected/fe4d6d85-749d-4ca8-a962-e49678672538-kube-api-access-8dmm8\") pod \"fe4d6d85-749d-4ca8-a962-e49678672538\" (UID: \"fe4d6d85-749d-4ca8-a962-e49678672538\") " Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.075960 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-ovsdbserver-nb\") pod \"fe4d6d85-749d-4ca8-a962-e49678672538\" (UID: \"fe4d6d85-749d-4ca8-a962-e49678672538\") " Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.084536 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe4d6d85-749d-4ca8-a962-e49678672538-kube-api-access-8dmm8" (OuterVolumeSpecName: "kube-api-access-8dmm8") pod "fe4d6d85-749d-4ca8-a962-e49678672538" (UID: "fe4d6d85-749d-4ca8-a962-e49678672538"). InnerVolumeSpecName "kube-api-access-8dmm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.128607 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fe4d6d85-749d-4ca8-a962-e49678672538" (UID: "fe4d6d85-749d-4ca8-a962-e49678672538"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.132381 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fe4d6d85-749d-4ca8-a962-e49678672538" (UID: "fe4d6d85-749d-4ca8-a962-e49678672538"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.139364 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fe4d6d85-749d-4ca8-a962-e49678672538" (UID: "fe4d6d85-749d-4ca8-a962-e49678672538"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.155521 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-config" (OuterVolumeSpecName: "config") pod "fe4d6d85-749d-4ca8-a962-e49678672538" (UID: "fe4d6d85-749d-4ca8-a962-e49678672538"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.177684 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dmm8\" (UniqueName: \"kubernetes.io/projected/fe4d6d85-749d-4ca8-a962-e49678672538-kube-api-access-8dmm8\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.177719 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.177731 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.177743 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.177755 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe4d6d85-749d-4ca8-a962-e49678672538-config\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.360142 4808 generic.go:334] "Generic (PLEG): container finished" podID="fe4d6d85-749d-4ca8-a962-e49678672538" containerID="14265a8079d6478a47893642811ce11637483a21a6d50eb089c39f67ea514e48" exitCode=0 Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.361105 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-85kbf" Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.366286 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-85kbf" event={"ID":"fe4d6d85-749d-4ca8-a962-e49678672538","Type":"ContainerDied","Data":"14265a8079d6478a47893642811ce11637483a21a6d50eb089c39f67ea514e48"} Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.366458 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-85kbf" event={"ID":"fe4d6d85-749d-4ca8-a962-e49678672538","Type":"ContainerDied","Data":"f00f248e9e91c2f987a5aec375c0ce40392e5d5d4ac539b2febbc43f32487e0e"} Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.366528 4808 scope.go:117] "RemoveContainer" containerID="14265a8079d6478a47893642811ce11637483a21a6d50eb089c39f67ea514e48" Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.417342 4808 scope.go:117] "RemoveContainer" containerID="6939b110451bfa0fdeb8fd14f58f6a84cfc2ce05af2949ca6015d2431e55fe8f" Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.425939 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-85kbf"] Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.426300 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-85kbf"] Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.440659 4808 scope.go:117] "RemoveContainer" containerID="14265a8079d6478a47893642811ce11637483a21a6d50eb089c39f67ea514e48" Oct 02 17:00:33 crc kubenswrapper[4808]: E1002 17:00:33.441089 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14265a8079d6478a47893642811ce11637483a21a6d50eb089c39f67ea514e48\": container with ID starting with 14265a8079d6478a47893642811ce11637483a21a6d50eb089c39f67ea514e48 not found: ID does not exist" containerID="14265a8079d6478a47893642811ce11637483a21a6d50eb089c39f67ea514e48" Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.441130 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14265a8079d6478a47893642811ce11637483a21a6d50eb089c39f67ea514e48"} err="failed to get container status \"14265a8079d6478a47893642811ce11637483a21a6d50eb089c39f67ea514e48\": rpc error: code = NotFound desc = could not find container \"14265a8079d6478a47893642811ce11637483a21a6d50eb089c39f67ea514e48\": container with ID starting with 14265a8079d6478a47893642811ce11637483a21a6d50eb089c39f67ea514e48 not found: ID does not exist" Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.441158 4808 scope.go:117] "RemoveContainer" containerID="6939b110451bfa0fdeb8fd14f58f6a84cfc2ce05af2949ca6015d2431e55fe8f" Oct 02 17:00:33 crc kubenswrapper[4808]: E1002 17:00:33.441617 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6939b110451bfa0fdeb8fd14f58f6a84cfc2ce05af2949ca6015d2431e55fe8f\": container with ID starting with 6939b110451bfa0fdeb8fd14f58f6a84cfc2ce05af2949ca6015d2431e55fe8f not found: ID does not exist" containerID="6939b110451bfa0fdeb8fd14f58f6a84cfc2ce05af2949ca6015d2431e55fe8f" Oct 02 17:00:33 crc kubenswrapper[4808]: I1002 17:00:33.441641 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6939b110451bfa0fdeb8fd14f58f6a84cfc2ce05af2949ca6015d2431e55fe8f"} err="failed to get container status \"6939b110451bfa0fdeb8fd14f58f6a84cfc2ce05af2949ca6015d2431e55fe8f\": rpc error: code = NotFound desc = could not find container \"6939b110451bfa0fdeb8fd14f58f6a84cfc2ce05af2949ca6015d2431e55fe8f\": container with ID starting with 6939b110451bfa0fdeb8fd14f58f6a84cfc2ce05af2949ca6015d2431e55fe8f not found: ID does not exist" Oct 02 17:00:35 crc kubenswrapper[4808]: I1002 17:00:35.417008 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe4d6d85-749d-4ca8-a962-e49678672538" path="/var/lib/kubelet/pods/fe4d6d85-749d-4ca8-a962-e49678672538/volumes" Oct 02 17:00:36 crc kubenswrapper[4808]: I1002 17:00:36.397183 4808 generic.go:334] "Generic (PLEG): container finished" podID="802da733-8c7c-450b-a9c6-7fda3f2d2fbc" containerID="9a65749ab5db1a8196465d4bae65ab12cec790701140fd3591a447236d973492" exitCode=0 Oct 02 17:00:36 crc kubenswrapper[4808]: I1002 17:00:36.397263 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-jkwct" event={"ID":"802da733-8c7c-450b-a9c6-7fda3f2d2fbc","Type":"ContainerDied","Data":"9a65749ab5db1a8196465d4bae65ab12cec790701140fd3591a447236d973492"} Oct 02 17:00:37 crc kubenswrapper[4808]: I1002 17:00:37.500689 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 17:00:37 crc kubenswrapper[4808]: I1002 17:00:37.500990 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" containerName="ceilometer-central-agent" containerID="cri-o://0f6479ff535d02280faff7cca75dffddfd4ae0daef90502722fe17ad3ffaa3fb" gracePeriod=30 Oct 02 17:00:37 crc kubenswrapper[4808]: I1002 17:00:37.501479 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" containerName="proxy-httpd" containerID="cri-o://95e245e2a282a68d227f87f0a4202fabdd57814fc3e943d366c73a22d098be10" gracePeriod=30 Oct 02 17:00:37 crc kubenswrapper[4808]: I1002 17:00:37.501541 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" containerName="sg-core" containerID="cri-o://c097f25e6215428c150710efb5bf442f8649de49ea19162ee15380e03db6bfbd" gracePeriod=30 Oct 02 17:00:37 crc kubenswrapper[4808]: I1002 17:00:37.501587 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" containerName="ceilometer-notification-agent" containerID="cri-o://0c3523d23eea77f3a26989699d6139b4952958e86d6780f8cad3f7bce59f1a19" gracePeriod=30 Oct 02 17:00:37 crc kubenswrapper[4808]: I1002 17:00:37.746760 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-jkwct" Oct 02 17:00:37 crc kubenswrapper[4808]: I1002 17:00:37.864883 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpt6n\" (UniqueName: \"kubernetes.io/projected/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-kube-api-access-zpt6n\") pod \"802da733-8c7c-450b-a9c6-7fda3f2d2fbc\" (UID: \"802da733-8c7c-450b-a9c6-7fda3f2d2fbc\") " Oct 02 17:00:37 crc kubenswrapper[4808]: I1002 17:00:37.864955 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-config-data\") pod \"802da733-8c7c-450b-a9c6-7fda3f2d2fbc\" (UID: \"802da733-8c7c-450b-a9c6-7fda3f2d2fbc\") " Oct 02 17:00:37 crc kubenswrapper[4808]: I1002 17:00:37.865039 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-combined-ca-bundle\") pod \"802da733-8c7c-450b-a9c6-7fda3f2d2fbc\" (UID: \"802da733-8c7c-450b-a9c6-7fda3f2d2fbc\") " Oct 02 17:00:37 crc kubenswrapper[4808]: I1002 17:00:37.865294 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-scripts\") pod \"802da733-8c7c-450b-a9c6-7fda3f2d2fbc\" (UID: \"802da733-8c7c-450b-a9c6-7fda3f2d2fbc\") " Oct 02 17:00:37 crc kubenswrapper[4808]: I1002 17:00:37.870838 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-kube-api-access-zpt6n" (OuterVolumeSpecName: "kube-api-access-zpt6n") pod "802da733-8c7c-450b-a9c6-7fda3f2d2fbc" (UID: "802da733-8c7c-450b-a9c6-7fda3f2d2fbc"). InnerVolumeSpecName "kube-api-access-zpt6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:00:37 crc kubenswrapper[4808]: I1002 17:00:37.887046 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-scripts" (OuterVolumeSpecName: "scripts") pod "802da733-8c7c-450b-a9c6-7fda3f2d2fbc" (UID: "802da733-8c7c-450b-a9c6-7fda3f2d2fbc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:37 crc kubenswrapper[4808]: I1002 17:00:37.891813 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "802da733-8c7c-450b-a9c6-7fda3f2d2fbc" (UID: "802da733-8c7c-450b-a9c6-7fda3f2d2fbc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:37 crc kubenswrapper[4808]: I1002 17:00:37.896662 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-config-data" (OuterVolumeSpecName: "config-data") pod "802da733-8c7c-450b-a9c6-7fda3f2d2fbc" (UID: "802da733-8c7c-450b-a9c6-7fda3f2d2fbc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:37 crc kubenswrapper[4808]: I1002 17:00:37.967065 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpt6n\" (UniqueName: \"kubernetes.io/projected/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-kube-api-access-zpt6n\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:37 crc kubenswrapper[4808]: I1002 17:00:37.967097 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:37 crc kubenswrapper[4808]: I1002 17:00:37.967106 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:37 crc kubenswrapper[4808]: I1002 17:00:37.967116 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/802da733-8c7c-450b-a9c6-7fda3f2d2fbc-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:38 crc kubenswrapper[4808]: I1002 17:00:38.428935 4808 generic.go:334] "Generic (PLEG): container finished" podID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" containerID="95e245e2a282a68d227f87f0a4202fabdd57814fc3e943d366c73a22d098be10" exitCode=0 Oct 02 17:00:38 crc kubenswrapper[4808]: I1002 17:00:38.429295 4808 generic.go:334] "Generic (PLEG): container finished" podID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" containerID="c097f25e6215428c150710efb5bf442f8649de49ea19162ee15380e03db6bfbd" exitCode=2 Oct 02 17:00:38 crc kubenswrapper[4808]: I1002 17:00:38.429410 4808 generic.go:334] "Generic (PLEG): container finished" podID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" containerID="0f6479ff535d02280faff7cca75dffddfd4ae0daef90502722fe17ad3ffaa3fb" exitCode=0 Oct 02 17:00:38 crc kubenswrapper[4808]: I1002 17:00:38.429024 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c66f29c7-9e93-4417-bc3e-b486a0b167cc","Type":"ContainerDied","Data":"95e245e2a282a68d227f87f0a4202fabdd57814fc3e943d366c73a22d098be10"} Oct 02 17:00:38 crc kubenswrapper[4808]: I1002 17:00:38.429674 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c66f29c7-9e93-4417-bc3e-b486a0b167cc","Type":"ContainerDied","Data":"c097f25e6215428c150710efb5bf442f8649de49ea19162ee15380e03db6bfbd"} Oct 02 17:00:38 crc kubenswrapper[4808]: I1002 17:00:38.429783 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c66f29c7-9e93-4417-bc3e-b486a0b167cc","Type":"ContainerDied","Data":"0f6479ff535d02280faff7cca75dffddfd4ae0daef90502722fe17ad3ffaa3fb"} Oct 02 17:00:38 crc kubenswrapper[4808]: I1002 17:00:38.431418 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-jkwct" event={"ID":"802da733-8c7c-450b-a9c6-7fda3f2d2fbc","Type":"ContainerDied","Data":"cb9778d80c516318bf9db29e486337092274f03159b04c42de8b98505bed1e98"} Oct 02 17:00:38 crc kubenswrapper[4808]: I1002 17:00:38.431456 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb9778d80c516318bf9db29e486337092274f03159b04c42de8b98505bed1e98" Oct 02 17:00:38 crc kubenswrapper[4808]: I1002 17:00:38.431638 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-jkwct" Oct 02 17:00:38 crc kubenswrapper[4808]: I1002 17:00:38.601418 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 17:00:38 crc kubenswrapper[4808]: I1002 17:00:38.601651 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c8a3b077-ac37-4c23-ad69-2916a00ce4b5" containerName="nova-api-log" containerID="cri-o://c3af4d6aaf63056ce1e2dd7d19db4405a1c4ff06857cb2311174c96ce9129041" gracePeriod=30 Oct 02 17:00:38 crc kubenswrapper[4808]: I1002 17:00:38.601756 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c8a3b077-ac37-4c23-ad69-2916a00ce4b5" containerName="nova-api-api" containerID="cri-o://557c30fafca450d2bb3f4df11db3e147992fe0f858ac738021c96fe43f85feb2" gracePeriod=30 Oct 02 17:00:38 crc kubenswrapper[4808]: I1002 17:00:38.624499 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 17:00:38 crc kubenswrapper[4808]: I1002 17:00:38.624977 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4aa280f6-f14f-47de-bbba-8f49229e4ec4" containerName="nova-scheduler-scheduler" containerID="cri-o://01261df500f04035637e93506f308f1fcd941eefda1095f190b78293ce07ac1e" gracePeriod=30 Oct 02 17:00:38 crc kubenswrapper[4808]: I1002 17:00:38.650930 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 17:00:38 crc kubenswrapper[4808]: I1002 17:00:38.651205 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8d7a4dd0-7af5-4d01-b888-14ce54434b2d" containerName="nova-metadata-log" containerID="cri-o://f49fb6be9af659916b2b1aa693d686204de7fd4bfc4f8c347de9f151a61c84d8" gracePeriod=30 Oct 02 17:00:38 crc kubenswrapper[4808]: I1002 17:00:38.651287 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8d7a4dd0-7af5-4d01-b888-14ce54434b2d" containerName="nova-metadata-metadata" containerID="cri-o://41cb07534ca1da05f1b1ba049073f435695baf47a1d14f6e92fbba1ee4be4032" gracePeriod=30 Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.220737 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.387917 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qd9jg\" (UniqueName: \"kubernetes.io/projected/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-kube-api-access-qd9jg\") pod \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.388038 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-internal-tls-certs\") pod \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.388072 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-combined-ca-bundle\") pod \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.388217 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-logs\") pod \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.388594 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-logs" (OuterVolumeSpecName: "logs") pod "c8a3b077-ac37-4c23-ad69-2916a00ce4b5" (UID: "c8a3b077-ac37-4c23-ad69-2916a00ce4b5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.388653 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-public-tls-certs\") pod \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.388919 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-config-data\") pod \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\" (UID: \"c8a3b077-ac37-4c23-ad69-2916a00ce4b5\") " Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.389579 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-logs\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.396359 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-kube-api-access-qd9jg" (OuterVolumeSpecName: "kube-api-access-qd9jg") pod "c8a3b077-ac37-4c23-ad69-2916a00ce4b5" (UID: "c8a3b077-ac37-4c23-ad69-2916a00ce4b5"). InnerVolumeSpecName "kube-api-access-qd9jg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.429444 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-config-data" (OuterVolumeSpecName: "config-data") pod "c8a3b077-ac37-4c23-ad69-2916a00ce4b5" (UID: "c8a3b077-ac37-4c23-ad69-2916a00ce4b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.429953 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8a3b077-ac37-4c23-ad69-2916a00ce4b5" (UID: "c8a3b077-ac37-4c23-ad69-2916a00ce4b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.454133 4808 generic.go:334] "Generic (PLEG): container finished" podID="c8a3b077-ac37-4c23-ad69-2916a00ce4b5" containerID="557c30fafca450d2bb3f4df11db3e147992fe0f858ac738021c96fe43f85feb2" exitCode=0 Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.454178 4808 generic.go:334] "Generic (PLEG): container finished" podID="c8a3b077-ac37-4c23-ad69-2916a00ce4b5" containerID="c3af4d6aaf63056ce1e2dd7d19db4405a1c4ff06857cb2311174c96ce9129041" exitCode=143 Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.454396 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.456922 4808 generic.go:334] "Generic (PLEG): container finished" podID="8d7a4dd0-7af5-4d01-b888-14ce54434b2d" containerID="f49fb6be9af659916b2b1aa693d686204de7fd4bfc4f8c347de9f151a61c84d8" exitCode=143 Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.465617 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c8a3b077-ac37-4c23-ad69-2916a00ce4b5" (UID: "c8a3b077-ac37-4c23-ad69-2916a00ce4b5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.465729 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c8a3b077-ac37-4c23-ad69-2916a00ce4b5" (UID: "c8a3b077-ac37-4c23-ad69-2916a00ce4b5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.465771 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c8a3b077-ac37-4c23-ad69-2916a00ce4b5","Type":"ContainerDied","Data":"557c30fafca450d2bb3f4df11db3e147992fe0f858ac738021c96fe43f85feb2"} Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.465809 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c8a3b077-ac37-4c23-ad69-2916a00ce4b5","Type":"ContainerDied","Data":"c3af4d6aaf63056ce1e2dd7d19db4405a1c4ff06857cb2311174c96ce9129041"} Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.465824 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c8a3b077-ac37-4c23-ad69-2916a00ce4b5","Type":"ContainerDied","Data":"201a7428185006e7ab205cff2f9cb249d373ee597bd924728bcef649857974bd"} Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.465837 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d7a4dd0-7af5-4d01-b888-14ce54434b2d","Type":"ContainerDied","Data":"f49fb6be9af659916b2b1aa693d686204de7fd4bfc4f8c347de9f151a61c84d8"} Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.465872 4808 scope.go:117] "RemoveContainer" containerID="557c30fafca450d2bb3f4df11db3e147992fe0f858ac738021c96fe43f85feb2" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.491139 4808 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.491190 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.491204 4808 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.491217 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.491245 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qd9jg\" (UniqueName: \"kubernetes.io/projected/c8a3b077-ac37-4c23-ad69-2916a00ce4b5-kube-api-access-qd9jg\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.502108 4808 scope.go:117] "RemoveContainer" containerID="c3af4d6aaf63056ce1e2dd7d19db4405a1c4ff06857cb2311174c96ce9129041" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.531104 4808 scope.go:117] "RemoveContainer" containerID="557c30fafca450d2bb3f4df11db3e147992fe0f858ac738021c96fe43f85feb2" Oct 02 17:00:39 crc kubenswrapper[4808]: E1002 17:00:39.532906 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"557c30fafca450d2bb3f4df11db3e147992fe0f858ac738021c96fe43f85feb2\": container with ID starting with 557c30fafca450d2bb3f4df11db3e147992fe0f858ac738021c96fe43f85feb2 not found: ID does not exist" containerID="557c30fafca450d2bb3f4df11db3e147992fe0f858ac738021c96fe43f85feb2" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.532955 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"557c30fafca450d2bb3f4df11db3e147992fe0f858ac738021c96fe43f85feb2"} err="failed to get container status \"557c30fafca450d2bb3f4df11db3e147992fe0f858ac738021c96fe43f85feb2\": rpc error: code = NotFound desc = could not find container \"557c30fafca450d2bb3f4df11db3e147992fe0f858ac738021c96fe43f85feb2\": container with ID starting with 557c30fafca450d2bb3f4df11db3e147992fe0f858ac738021c96fe43f85feb2 not found: ID does not exist" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.532988 4808 scope.go:117] "RemoveContainer" containerID="c3af4d6aaf63056ce1e2dd7d19db4405a1c4ff06857cb2311174c96ce9129041" Oct 02 17:00:39 crc kubenswrapper[4808]: E1002 17:00:39.533394 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3af4d6aaf63056ce1e2dd7d19db4405a1c4ff06857cb2311174c96ce9129041\": container with ID starting with c3af4d6aaf63056ce1e2dd7d19db4405a1c4ff06857cb2311174c96ce9129041 not found: ID does not exist" containerID="c3af4d6aaf63056ce1e2dd7d19db4405a1c4ff06857cb2311174c96ce9129041" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.533421 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3af4d6aaf63056ce1e2dd7d19db4405a1c4ff06857cb2311174c96ce9129041"} err="failed to get container status \"c3af4d6aaf63056ce1e2dd7d19db4405a1c4ff06857cb2311174c96ce9129041\": rpc error: code = NotFound desc = could not find container \"c3af4d6aaf63056ce1e2dd7d19db4405a1c4ff06857cb2311174c96ce9129041\": container with ID starting with c3af4d6aaf63056ce1e2dd7d19db4405a1c4ff06857cb2311174c96ce9129041 not found: ID does not exist" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.533436 4808 scope.go:117] "RemoveContainer" containerID="557c30fafca450d2bb3f4df11db3e147992fe0f858ac738021c96fe43f85feb2" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.533794 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"557c30fafca450d2bb3f4df11db3e147992fe0f858ac738021c96fe43f85feb2"} err="failed to get container status \"557c30fafca450d2bb3f4df11db3e147992fe0f858ac738021c96fe43f85feb2\": rpc error: code = NotFound desc = could not find container \"557c30fafca450d2bb3f4df11db3e147992fe0f858ac738021c96fe43f85feb2\": container with ID starting with 557c30fafca450d2bb3f4df11db3e147992fe0f858ac738021c96fe43f85feb2 not found: ID does not exist" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.533813 4808 scope.go:117] "RemoveContainer" containerID="c3af4d6aaf63056ce1e2dd7d19db4405a1c4ff06857cb2311174c96ce9129041" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.534076 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3af4d6aaf63056ce1e2dd7d19db4405a1c4ff06857cb2311174c96ce9129041"} err="failed to get container status \"c3af4d6aaf63056ce1e2dd7d19db4405a1c4ff06857cb2311174c96ce9129041\": rpc error: code = NotFound desc = could not find container \"c3af4d6aaf63056ce1e2dd7d19db4405a1c4ff06857cb2311174c96ce9129041\": container with ID starting with c3af4d6aaf63056ce1e2dd7d19db4405a1c4ff06857cb2311174c96ce9129041 not found: ID does not exist" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.817558 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.854832 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.864877 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.877084 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 17:00:39 crc kubenswrapper[4808]: E1002 17:00:39.878360 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="802da733-8c7c-450b-a9c6-7fda3f2d2fbc" containerName="nova-manage" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.878378 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="802da733-8c7c-450b-a9c6-7fda3f2d2fbc" containerName="nova-manage" Oct 02 17:00:39 crc kubenswrapper[4808]: E1002 17:00:39.878391 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe4d6d85-749d-4ca8-a962-e49678672538" containerName="init" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.878396 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe4d6d85-749d-4ca8-a962-e49678672538" containerName="init" Oct 02 17:00:39 crc kubenswrapper[4808]: E1002 17:00:39.878411 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe4d6d85-749d-4ca8-a962-e49678672538" containerName="dnsmasq-dns" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.878417 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe4d6d85-749d-4ca8-a962-e49678672538" containerName="dnsmasq-dns" Oct 02 17:00:39 crc kubenswrapper[4808]: E1002 17:00:39.878435 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8a3b077-ac37-4c23-ad69-2916a00ce4b5" containerName="nova-api-log" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.878440 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8a3b077-ac37-4c23-ad69-2916a00ce4b5" containerName="nova-api-log" Oct 02 17:00:39 crc kubenswrapper[4808]: E1002 17:00:39.878449 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8a3b077-ac37-4c23-ad69-2916a00ce4b5" containerName="nova-api-api" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.878475 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8a3b077-ac37-4c23-ad69-2916a00ce4b5" containerName="nova-api-api" Oct 02 17:00:39 crc kubenswrapper[4808]: E1002 17:00:39.878490 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa280f6-f14f-47de-bbba-8f49229e4ec4" containerName="nova-scheduler-scheduler" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.878497 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa280f6-f14f-47de-bbba-8f49229e4ec4" containerName="nova-scheduler-scheduler" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.878653 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="802da733-8c7c-450b-a9c6-7fda3f2d2fbc" containerName="nova-manage" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.878669 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8a3b077-ac37-4c23-ad69-2916a00ce4b5" containerName="nova-api-log" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.878681 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8a3b077-ac37-4c23-ad69-2916a00ce4b5" containerName="nova-api-api" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.878691 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe4d6d85-749d-4ca8-a962-e49678672538" containerName="dnsmasq-dns" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.878701 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aa280f6-f14f-47de-bbba-8f49229e4ec4" containerName="nova-scheduler-scheduler" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.879675 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.881400 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.884846 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.886292 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.889648 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.902699 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa5606b5-d0de-4403-b815-41d3216754d5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fa5606b5-d0de-4403-b815-41d3216754d5\") " pod="openstack/nova-api-0" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.902932 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa5606b5-d0de-4403-b815-41d3216754d5-logs\") pod \"nova-api-0\" (UID: \"fa5606b5-d0de-4403-b815-41d3216754d5\") " pod="openstack/nova-api-0" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.903118 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa5606b5-d0de-4403-b815-41d3216754d5-config-data\") pod \"nova-api-0\" (UID: \"fa5606b5-d0de-4403-b815-41d3216754d5\") " pod="openstack/nova-api-0" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.903346 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp6nx\" (UniqueName: \"kubernetes.io/projected/fa5606b5-d0de-4403-b815-41d3216754d5-kube-api-access-vp6nx\") pod \"nova-api-0\" (UID: \"fa5606b5-d0de-4403-b815-41d3216754d5\") " pod="openstack/nova-api-0" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.903485 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa5606b5-d0de-4403-b815-41d3216754d5-public-tls-certs\") pod \"nova-api-0\" (UID: \"fa5606b5-d0de-4403-b815-41d3216754d5\") " pod="openstack/nova-api-0" Oct 02 17:00:39 crc kubenswrapper[4808]: I1002 17:00:39.903591 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa5606b5-d0de-4403-b815-41d3216754d5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"fa5606b5-d0de-4403-b815-41d3216754d5\") " pod="openstack/nova-api-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.004264 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmrh4\" (UniqueName: \"kubernetes.io/projected/4aa280f6-f14f-47de-bbba-8f49229e4ec4-kube-api-access-zmrh4\") pod \"4aa280f6-f14f-47de-bbba-8f49229e4ec4\" (UID: \"4aa280f6-f14f-47de-bbba-8f49229e4ec4\") " Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.004428 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa280f6-f14f-47de-bbba-8f49229e4ec4-config-data\") pod \"4aa280f6-f14f-47de-bbba-8f49229e4ec4\" (UID: \"4aa280f6-f14f-47de-bbba-8f49229e4ec4\") " Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.004465 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa280f6-f14f-47de-bbba-8f49229e4ec4-combined-ca-bundle\") pod \"4aa280f6-f14f-47de-bbba-8f49229e4ec4\" (UID: \"4aa280f6-f14f-47de-bbba-8f49229e4ec4\") " Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.004689 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa5606b5-d0de-4403-b815-41d3216754d5-public-tls-certs\") pod \"nova-api-0\" (UID: \"fa5606b5-d0de-4403-b815-41d3216754d5\") " pod="openstack/nova-api-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.004734 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa5606b5-d0de-4403-b815-41d3216754d5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"fa5606b5-d0de-4403-b815-41d3216754d5\") " pod="openstack/nova-api-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.004763 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa5606b5-d0de-4403-b815-41d3216754d5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fa5606b5-d0de-4403-b815-41d3216754d5\") " pod="openstack/nova-api-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.004811 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa5606b5-d0de-4403-b815-41d3216754d5-logs\") pod \"nova-api-0\" (UID: \"fa5606b5-d0de-4403-b815-41d3216754d5\") " pod="openstack/nova-api-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.004850 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa5606b5-d0de-4403-b815-41d3216754d5-config-data\") pod \"nova-api-0\" (UID: \"fa5606b5-d0de-4403-b815-41d3216754d5\") " pod="openstack/nova-api-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.004898 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp6nx\" (UniqueName: \"kubernetes.io/projected/fa5606b5-d0de-4403-b815-41d3216754d5-kube-api-access-vp6nx\") pod \"nova-api-0\" (UID: \"fa5606b5-d0de-4403-b815-41d3216754d5\") " pod="openstack/nova-api-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.005313 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa5606b5-d0de-4403-b815-41d3216754d5-logs\") pod \"nova-api-0\" (UID: \"fa5606b5-d0de-4403-b815-41d3216754d5\") " pod="openstack/nova-api-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.008041 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aa280f6-f14f-47de-bbba-8f49229e4ec4-kube-api-access-zmrh4" (OuterVolumeSpecName: "kube-api-access-zmrh4") pod "4aa280f6-f14f-47de-bbba-8f49229e4ec4" (UID: "4aa280f6-f14f-47de-bbba-8f49229e4ec4"). InnerVolumeSpecName "kube-api-access-zmrh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.008437 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa5606b5-d0de-4403-b815-41d3216754d5-public-tls-certs\") pod \"nova-api-0\" (UID: \"fa5606b5-d0de-4403-b815-41d3216754d5\") " pod="openstack/nova-api-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.009568 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa5606b5-d0de-4403-b815-41d3216754d5-config-data\") pod \"nova-api-0\" (UID: \"fa5606b5-d0de-4403-b815-41d3216754d5\") " pod="openstack/nova-api-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.009893 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa5606b5-d0de-4403-b815-41d3216754d5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fa5606b5-d0de-4403-b815-41d3216754d5\") " pod="openstack/nova-api-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.009943 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa5606b5-d0de-4403-b815-41d3216754d5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"fa5606b5-d0de-4403-b815-41d3216754d5\") " pod="openstack/nova-api-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.021594 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp6nx\" (UniqueName: \"kubernetes.io/projected/fa5606b5-d0de-4403-b815-41d3216754d5-kube-api-access-vp6nx\") pod \"nova-api-0\" (UID: \"fa5606b5-d0de-4403-b815-41d3216754d5\") " pod="openstack/nova-api-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.041636 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aa280f6-f14f-47de-bbba-8f49229e4ec4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4aa280f6-f14f-47de-bbba-8f49229e4ec4" (UID: "4aa280f6-f14f-47de-bbba-8f49229e4ec4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.045424 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aa280f6-f14f-47de-bbba-8f49229e4ec4-config-data" (OuterVolumeSpecName: "config-data") pod "4aa280f6-f14f-47de-bbba-8f49229e4ec4" (UID: "4aa280f6-f14f-47de-bbba-8f49229e4ec4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.106028 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmrh4\" (UniqueName: \"kubernetes.io/projected/4aa280f6-f14f-47de-bbba-8f49229e4ec4-kube-api-access-zmrh4\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.106057 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa280f6-f14f-47de-bbba-8f49229e4ec4-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.106069 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa280f6-f14f-47de-bbba-8f49229e4ec4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.221634 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.466554 4808 generic.go:334] "Generic (PLEG): container finished" podID="4aa280f6-f14f-47de-bbba-8f49229e4ec4" containerID="01261df500f04035637e93506f308f1fcd941eefda1095f190b78293ce07ac1e" exitCode=0 Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.466613 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.466632 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4aa280f6-f14f-47de-bbba-8f49229e4ec4","Type":"ContainerDied","Data":"01261df500f04035637e93506f308f1fcd941eefda1095f190b78293ce07ac1e"} Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.466663 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4aa280f6-f14f-47de-bbba-8f49229e4ec4","Type":"ContainerDied","Data":"222e6c947b2832af5877014b9ee54911ef4805eeea1f45c1c3f6699210ee4180"} Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.466685 4808 scope.go:117] "RemoveContainer" containerID="01261df500f04035637e93506f308f1fcd941eefda1095f190b78293ce07ac1e" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.507030 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.521992 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.527459 4808 scope.go:117] "RemoveContainer" containerID="01261df500f04035637e93506f308f1fcd941eefda1095f190b78293ce07ac1e" Oct 02 17:00:40 crc kubenswrapper[4808]: E1002 17:00:40.527947 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01261df500f04035637e93506f308f1fcd941eefda1095f190b78293ce07ac1e\": container with ID starting with 01261df500f04035637e93506f308f1fcd941eefda1095f190b78293ce07ac1e not found: ID does not exist" containerID="01261df500f04035637e93506f308f1fcd941eefda1095f190b78293ce07ac1e" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.527984 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01261df500f04035637e93506f308f1fcd941eefda1095f190b78293ce07ac1e"} err="failed to get container status \"01261df500f04035637e93506f308f1fcd941eefda1095f190b78293ce07ac1e\": rpc error: code = NotFound desc = could not find container \"01261df500f04035637e93506f308f1fcd941eefda1095f190b78293ce07ac1e\": container with ID starting with 01261df500f04035637e93506f308f1fcd941eefda1095f190b78293ce07ac1e not found: ID does not exist" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.542318 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.543477 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.546688 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.550074 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.682502 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 17:00:40 crc kubenswrapper[4808]: W1002 17:00:40.692804 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa5606b5_d0de_4403_b815_41d3216754d5.slice/crio-37c827046d15a7b09c3f4e7cbac6bbfb440ff62477de03e29411454bf42a7e2f WatchSource:0}: Error finding container 37c827046d15a7b09c3f4e7cbac6bbfb440ff62477de03e29411454bf42a7e2f: Status 404 returned error can't find the container with id 37c827046d15a7b09c3f4e7cbac6bbfb440ff62477de03e29411454bf42a7e2f Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.717315 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cebaac7-db1e-4020-ab4a-b63ebf95124d-config-data\") pod \"nova-scheduler-0\" (UID: \"1cebaac7-db1e-4020-ab4a-b63ebf95124d\") " pod="openstack/nova-scheduler-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.717432 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bvqz\" (UniqueName: \"kubernetes.io/projected/1cebaac7-db1e-4020-ab4a-b63ebf95124d-kube-api-access-5bvqz\") pod \"nova-scheduler-0\" (UID: \"1cebaac7-db1e-4020-ab4a-b63ebf95124d\") " pod="openstack/nova-scheduler-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.717491 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cebaac7-db1e-4020-ab4a-b63ebf95124d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1cebaac7-db1e-4020-ab4a-b63ebf95124d\") " pod="openstack/nova-scheduler-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.818752 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cebaac7-db1e-4020-ab4a-b63ebf95124d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1cebaac7-db1e-4020-ab4a-b63ebf95124d\") " pod="openstack/nova-scheduler-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.819546 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cebaac7-db1e-4020-ab4a-b63ebf95124d-config-data\") pod \"nova-scheduler-0\" (UID: \"1cebaac7-db1e-4020-ab4a-b63ebf95124d\") " pod="openstack/nova-scheduler-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.819694 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bvqz\" (UniqueName: \"kubernetes.io/projected/1cebaac7-db1e-4020-ab4a-b63ebf95124d-kube-api-access-5bvqz\") pod \"nova-scheduler-0\" (UID: \"1cebaac7-db1e-4020-ab4a-b63ebf95124d\") " pod="openstack/nova-scheduler-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.825536 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cebaac7-db1e-4020-ab4a-b63ebf95124d-config-data\") pod \"nova-scheduler-0\" (UID: \"1cebaac7-db1e-4020-ab4a-b63ebf95124d\") " pod="openstack/nova-scheduler-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.825823 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cebaac7-db1e-4020-ab4a-b63ebf95124d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1cebaac7-db1e-4020-ab4a-b63ebf95124d\") " pod="openstack/nova-scheduler-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.836795 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bvqz\" (UniqueName: \"kubernetes.io/projected/1cebaac7-db1e-4020-ab4a-b63ebf95124d-kube-api-access-5bvqz\") pod \"nova-scheduler-0\" (UID: \"1cebaac7-db1e-4020-ab4a-b63ebf95124d\") " pod="openstack/nova-scheduler-0" Oct 02 17:00:40 crc kubenswrapper[4808]: I1002 17:00:40.866013 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 17:00:41 crc kubenswrapper[4808]: I1002 17:00:41.294139 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 17:00:41 crc kubenswrapper[4808]: I1002 17:00:41.410366 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4aa280f6-f14f-47de-bbba-8f49229e4ec4" path="/var/lib/kubelet/pods/4aa280f6-f14f-47de-bbba-8f49229e4ec4/volumes" Oct 02 17:00:41 crc kubenswrapper[4808]: I1002 17:00:41.411913 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8a3b077-ac37-4c23-ad69-2916a00ce4b5" path="/var/lib/kubelet/pods/c8a3b077-ac37-4c23-ad69-2916a00ce4b5/volumes" Oct 02 17:00:41 crc kubenswrapper[4808]: I1002 17:00:41.479073 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fa5606b5-d0de-4403-b815-41d3216754d5","Type":"ContainerStarted","Data":"742675f6b0f0c210e7c15eb32b7ed177d84e0d8d3690f3654275991f5a02f03c"} Oct 02 17:00:41 crc kubenswrapper[4808]: I1002 17:00:41.479111 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fa5606b5-d0de-4403-b815-41d3216754d5","Type":"ContainerStarted","Data":"4c44082bda031b0e50d4c99647ab59de52c9f9d919d36f2f3c1d5fdb9516e3f5"} Oct 02 17:00:41 crc kubenswrapper[4808]: I1002 17:00:41.479122 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fa5606b5-d0de-4403-b815-41d3216754d5","Type":"ContainerStarted","Data":"37c827046d15a7b09c3f4e7cbac6bbfb440ff62477de03e29411454bf42a7e2f"} Oct 02 17:00:41 crc kubenswrapper[4808]: I1002 17:00:41.480497 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1cebaac7-db1e-4020-ab4a-b63ebf95124d","Type":"ContainerStarted","Data":"0a7a8bdf65140f2c37cc8e726eb8787fecbbfb75f95356567e192282f2fc61d1"} Oct 02 17:00:41 crc kubenswrapper[4808]: I1002 17:00:41.506366 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.50632828 podStartE2EDuration="2.50632828s" podCreationTimestamp="2025-10-02 17:00:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 17:00:41.504617254 +0000 UTC m=+1208.830146254" watchObservedRunningTime="2025-10-02 17:00:41.50632828 +0000 UTC m=+1208.831857300" Oct 02 17:00:41 crc kubenswrapper[4808]: I1002 17:00:41.536944 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.536923329 podStartE2EDuration="1.536923329s" podCreationTimestamp="2025-10-02 17:00:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 17:00:41.532557152 +0000 UTC m=+1208.858086172" watchObservedRunningTime="2025-10-02 17:00:41.536923329 +0000 UTC m=+1208.862452349" Oct 02 17:00:41 crc kubenswrapper[4808]: I1002 17:00:41.789702 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="8d7a4dd0-7af5-4d01-b888-14ce54434b2d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.178:8775/\": read tcp 10.217.0.2:49956->10.217.0.178:8775: read: connection reset by peer" Oct 02 17:00:41 crc kubenswrapper[4808]: I1002 17:00:41.789701 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="8d7a4dd0-7af5-4d01-b888-14ce54434b2d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.178:8775/\": read tcp 10.217.0.2:49958->10.217.0.178:8775: read: connection reset by peer" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.282931 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.451005 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-nova-metadata-tls-certs\") pod \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\" (UID: \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\") " Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.451073 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-combined-ca-bundle\") pod \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\" (UID: \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\") " Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.451126 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-config-data\") pod \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\" (UID: \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\") " Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.451266 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-logs\") pod \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\" (UID: \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\") " Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.451319 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqtrf\" (UniqueName: \"kubernetes.io/projected/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-kube-api-access-vqtrf\") pod \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\" (UID: \"8d7a4dd0-7af5-4d01-b888-14ce54434b2d\") " Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.457192 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-logs" (OuterVolumeSpecName: "logs") pod "8d7a4dd0-7af5-4d01-b888-14ce54434b2d" (UID: "8d7a4dd0-7af5-4d01-b888-14ce54434b2d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.472540 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-kube-api-access-vqtrf" (OuterVolumeSpecName: "kube-api-access-vqtrf") pod "8d7a4dd0-7af5-4d01-b888-14ce54434b2d" (UID: "8d7a4dd0-7af5-4d01-b888-14ce54434b2d"). InnerVolumeSpecName "kube-api-access-vqtrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.483420 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-config-data" (OuterVolumeSpecName: "config-data") pod "8d7a4dd0-7af5-4d01-b888-14ce54434b2d" (UID: "8d7a4dd0-7af5-4d01-b888-14ce54434b2d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.483844 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d7a4dd0-7af5-4d01-b888-14ce54434b2d" (UID: "8d7a4dd0-7af5-4d01-b888-14ce54434b2d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.492211 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1cebaac7-db1e-4020-ab4a-b63ebf95124d","Type":"ContainerStarted","Data":"b5fe82eab86bf6fc5459d9a1d2b0ce3fc6e77ea33ff938f32530b789e243c30c"} Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.499825 4808 generic.go:334] "Generic (PLEG): container finished" podID="8d7a4dd0-7af5-4d01-b888-14ce54434b2d" containerID="41cb07534ca1da05f1b1ba049073f435695baf47a1d14f6e92fbba1ee4be4032" exitCode=0 Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.500160 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.502494 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d7a4dd0-7af5-4d01-b888-14ce54434b2d","Type":"ContainerDied","Data":"41cb07534ca1da05f1b1ba049073f435695baf47a1d14f6e92fbba1ee4be4032"} Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.502533 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d7a4dd0-7af5-4d01-b888-14ce54434b2d","Type":"ContainerDied","Data":"68650fb34154384ab162335c1acb367c0475844cbefa9e8eca63fd27af9aec52"} Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.502565 4808 scope.go:117] "RemoveContainer" containerID="41cb07534ca1da05f1b1ba049073f435695baf47a1d14f6e92fbba1ee4be4032" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.506421 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "8d7a4dd0-7af5-4d01-b888-14ce54434b2d" (UID: "8d7a4dd0-7af5-4d01-b888-14ce54434b2d"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.529466 4808 scope.go:117] "RemoveContainer" containerID="f49fb6be9af659916b2b1aa693d686204de7fd4bfc4f8c347de9f151a61c84d8" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.547744 4808 scope.go:117] "RemoveContainer" containerID="41cb07534ca1da05f1b1ba049073f435695baf47a1d14f6e92fbba1ee4be4032" Oct 02 17:00:42 crc kubenswrapper[4808]: E1002 17:00:42.548399 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41cb07534ca1da05f1b1ba049073f435695baf47a1d14f6e92fbba1ee4be4032\": container with ID starting with 41cb07534ca1da05f1b1ba049073f435695baf47a1d14f6e92fbba1ee4be4032 not found: ID does not exist" containerID="41cb07534ca1da05f1b1ba049073f435695baf47a1d14f6e92fbba1ee4be4032" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.548456 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41cb07534ca1da05f1b1ba049073f435695baf47a1d14f6e92fbba1ee4be4032"} err="failed to get container status \"41cb07534ca1da05f1b1ba049073f435695baf47a1d14f6e92fbba1ee4be4032\": rpc error: code = NotFound desc = could not find container \"41cb07534ca1da05f1b1ba049073f435695baf47a1d14f6e92fbba1ee4be4032\": container with ID starting with 41cb07534ca1da05f1b1ba049073f435695baf47a1d14f6e92fbba1ee4be4032 not found: ID does not exist" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.548506 4808 scope.go:117] "RemoveContainer" containerID="f49fb6be9af659916b2b1aa693d686204de7fd4bfc4f8c347de9f151a61c84d8" Oct 02 17:00:42 crc kubenswrapper[4808]: E1002 17:00:42.548857 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f49fb6be9af659916b2b1aa693d686204de7fd4bfc4f8c347de9f151a61c84d8\": container with ID starting with f49fb6be9af659916b2b1aa693d686204de7fd4bfc4f8c347de9f151a61c84d8 not found: ID does not exist" containerID="f49fb6be9af659916b2b1aa693d686204de7fd4bfc4f8c347de9f151a61c84d8" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.548901 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f49fb6be9af659916b2b1aa693d686204de7fd4bfc4f8c347de9f151a61c84d8"} err="failed to get container status \"f49fb6be9af659916b2b1aa693d686204de7fd4bfc4f8c347de9f151a61c84d8\": rpc error: code = NotFound desc = could not find container \"f49fb6be9af659916b2b1aa693d686204de7fd4bfc4f8c347de9f151a61c84d8\": container with ID starting with f49fb6be9af659916b2b1aa693d686204de7fd4bfc4f8c347de9f151a61c84d8 not found: ID does not exist" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.554010 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.554031 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-logs\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.554041 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqtrf\" (UniqueName: \"kubernetes.io/projected/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-kube-api-access-vqtrf\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.554051 4808 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.554059 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d7a4dd0-7af5-4d01-b888-14ce54434b2d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.844476 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.870621 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.882856 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 17:00:42 crc kubenswrapper[4808]: E1002 17:00:42.883291 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d7a4dd0-7af5-4d01-b888-14ce54434b2d" containerName="nova-metadata-log" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.883311 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d7a4dd0-7af5-4d01-b888-14ce54434b2d" containerName="nova-metadata-log" Oct 02 17:00:42 crc kubenswrapper[4808]: E1002 17:00:42.883343 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d7a4dd0-7af5-4d01-b888-14ce54434b2d" containerName="nova-metadata-metadata" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.883352 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d7a4dd0-7af5-4d01-b888-14ce54434b2d" containerName="nova-metadata-metadata" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.883574 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d7a4dd0-7af5-4d01-b888-14ce54434b2d" containerName="nova-metadata-log" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.883599 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d7a4dd0-7af5-4d01-b888-14ce54434b2d" containerName="nova-metadata-metadata" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.895903 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.896050 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.899761 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 17:00:42 crc kubenswrapper[4808]: I1002 17:00:42.900747 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 17:00:43 crc kubenswrapper[4808]: I1002 17:00:43.063691 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcs2b\" (UniqueName: \"kubernetes.io/projected/c0748c83-a144-4c4c-9980-41c6592edf0f-kube-api-access-hcs2b\") pod \"nova-metadata-0\" (UID: \"c0748c83-a144-4c4c-9980-41c6592edf0f\") " pod="openstack/nova-metadata-0" Oct 02 17:00:43 crc kubenswrapper[4808]: I1002 17:00:43.064515 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0748c83-a144-4c4c-9980-41c6592edf0f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c0748c83-a144-4c4c-9980-41c6592edf0f\") " pod="openstack/nova-metadata-0" Oct 02 17:00:43 crc kubenswrapper[4808]: I1002 17:00:43.064692 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0748c83-a144-4c4c-9980-41c6592edf0f-config-data\") pod \"nova-metadata-0\" (UID: \"c0748c83-a144-4c4c-9980-41c6592edf0f\") " pod="openstack/nova-metadata-0" Oct 02 17:00:43 crc kubenswrapper[4808]: I1002 17:00:43.064857 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0748c83-a144-4c4c-9980-41c6592edf0f-logs\") pod \"nova-metadata-0\" (UID: \"c0748c83-a144-4c4c-9980-41c6592edf0f\") " pod="openstack/nova-metadata-0" Oct 02 17:00:43 crc kubenswrapper[4808]: I1002 17:00:43.065031 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0748c83-a144-4c4c-9980-41c6592edf0f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c0748c83-a144-4c4c-9980-41c6592edf0f\") " pod="openstack/nova-metadata-0" Oct 02 17:00:43 crc kubenswrapper[4808]: I1002 17:00:43.167575 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcs2b\" (UniqueName: \"kubernetes.io/projected/c0748c83-a144-4c4c-9980-41c6592edf0f-kube-api-access-hcs2b\") pod \"nova-metadata-0\" (UID: \"c0748c83-a144-4c4c-9980-41c6592edf0f\") " pod="openstack/nova-metadata-0" Oct 02 17:00:43 crc kubenswrapper[4808]: I1002 17:00:43.168086 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0748c83-a144-4c4c-9980-41c6592edf0f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c0748c83-a144-4c4c-9980-41c6592edf0f\") " pod="openstack/nova-metadata-0" Oct 02 17:00:43 crc kubenswrapper[4808]: I1002 17:00:43.168310 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0748c83-a144-4c4c-9980-41c6592edf0f-config-data\") pod \"nova-metadata-0\" (UID: \"c0748c83-a144-4c4c-9980-41c6592edf0f\") " pod="openstack/nova-metadata-0" Oct 02 17:00:43 crc kubenswrapper[4808]: I1002 17:00:43.168494 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0748c83-a144-4c4c-9980-41c6592edf0f-logs\") pod \"nova-metadata-0\" (UID: \"c0748c83-a144-4c4c-9980-41c6592edf0f\") " pod="openstack/nova-metadata-0" Oct 02 17:00:43 crc kubenswrapper[4808]: I1002 17:00:43.168748 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0748c83-a144-4c4c-9980-41c6592edf0f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c0748c83-a144-4c4c-9980-41c6592edf0f\") " pod="openstack/nova-metadata-0" Oct 02 17:00:43 crc kubenswrapper[4808]: I1002 17:00:43.169137 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0748c83-a144-4c4c-9980-41c6592edf0f-logs\") pod \"nova-metadata-0\" (UID: \"c0748c83-a144-4c4c-9980-41c6592edf0f\") " pod="openstack/nova-metadata-0" Oct 02 17:00:43 crc kubenswrapper[4808]: I1002 17:00:43.173577 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0748c83-a144-4c4c-9980-41c6592edf0f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c0748c83-a144-4c4c-9980-41c6592edf0f\") " pod="openstack/nova-metadata-0" Oct 02 17:00:43 crc kubenswrapper[4808]: I1002 17:00:43.174020 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0748c83-a144-4c4c-9980-41c6592edf0f-config-data\") pod \"nova-metadata-0\" (UID: \"c0748c83-a144-4c4c-9980-41c6592edf0f\") " pod="openstack/nova-metadata-0" Oct 02 17:00:43 crc kubenswrapper[4808]: I1002 17:00:43.177806 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0748c83-a144-4c4c-9980-41c6592edf0f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c0748c83-a144-4c4c-9980-41c6592edf0f\") " pod="openstack/nova-metadata-0" Oct 02 17:00:43 crc kubenswrapper[4808]: I1002 17:00:43.191145 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcs2b\" (UniqueName: \"kubernetes.io/projected/c0748c83-a144-4c4c-9980-41c6592edf0f-kube-api-access-hcs2b\") pod \"nova-metadata-0\" (UID: \"c0748c83-a144-4c4c-9980-41c6592edf0f\") " pod="openstack/nova-metadata-0" Oct 02 17:00:43 crc kubenswrapper[4808]: I1002 17:00:43.224029 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 17:00:43 crc kubenswrapper[4808]: I1002 17:00:43.417251 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d7a4dd0-7af5-4d01-b888-14ce54434b2d" path="/var/lib/kubelet/pods/8d7a4dd0-7af5-4d01-b888-14ce54434b2d/volumes" Oct 02 17:00:43 crc kubenswrapper[4808]: I1002 17:00:43.690326 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 17:00:43 crc kubenswrapper[4808]: W1002 17:00:43.695752 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0748c83_a144_4c4c_9980_41c6592edf0f.slice/crio-8d18b1e184640e7d9d80c6f0ac9380d00d8515f6cfff91d2c6e54a0a8c66cb5e WatchSource:0}: Error finding container 8d18b1e184640e7d9d80c6f0ac9380d00d8515f6cfff91d2c6e54a0a8c66cb5e: Status 404 returned error can't find the container with id 8d18b1e184640e7d9d80c6f0ac9380d00d8515f6cfff91d2c6e54a0a8c66cb5e Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.410968 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.538222 4808 generic.go:334] "Generic (PLEG): container finished" podID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" containerID="0c3523d23eea77f3a26989699d6139b4952958e86d6780f8cad3f7bce59f1a19" exitCode=0 Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.538305 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c66f29c7-9e93-4417-bc3e-b486a0b167cc","Type":"ContainerDied","Data":"0c3523d23eea77f3a26989699d6139b4952958e86d6780f8cad3f7bce59f1a19"} Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.538336 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c66f29c7-9e93-4417-bc3e-b486a0b167cc","Type":"ContainerDied","Data":"76e577573bdd504776f07aac31b5ff3c20243b4ddb7d01494f3dcc9e7992e2d9"} Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.538356 4808 scope.go:117] "RemoveContainer" containerID="95e245e2a282a68d227f87f0a4202fabdd57814fc3e943d366c73a22d098be10" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.538558 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.540056 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c0748c83-a144-4c4c-9980-41c6592edf0f","Type":"ContainerStarted","Data":"4ed37a62dc22a8f703d50f5b70ab3959318448a914df5d841630c00864067be2"} Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.540083 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c0748c83-a144-4c4c-9980-41c6592edf0f","Type":"ContainerStarted","Data":"fed0ed468a7270b2990106c9846cd712dbd08556e18797640b72343f07ed1066"} Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.540095 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c0748c83-a144-4c4c-9980-41c6592edf0f","Type":"ContainerStarted","Data":"8d18b1e184640e7d9d80c6f0ac9380d00d8515f6cfff91d2c6e54a0a8c66cb5e"} Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.572012 4808 scope.go:117] "RemoveContainer" containerID="c097f25e6215428c150710efb5bf442f8649de49ea19162ee15380e03db6bfbd" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.573866 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.573854324 podStartE2EDuration="2.573854324s" podCreationTimestamp="2025-10-02 17:00:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 17:00:44.561100473 +0000 UTC m=+1211.886629483" watchObservedRunningTime="2025-10-02 17:00:44.573854324 +0000 UTC m=+1211.899383334" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.592519 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-combined-ca-bundle\") pod \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.592576 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-ceilometer-tls-certs\") pod \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.592601 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcnsx\" (UniqueName: \"kubernetes.io/projected/c66f29c7-9e93-4417-bc3e-b486a0b167cc-kube-api-access-wcnsx\") pod \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.592686 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-config-data\") pod \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.593365 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c66f29c7-9e93-4417-bc3e-b486a0b167cc-log-httpd\") pod \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.593477 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-scripts\") pod \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.593551 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c66f29c7-9e93-4417-bc3e-b486a0b167cc-run-httpd\") pod \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.593631 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-sg-core-conf-yaml\") pod \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\" (UID: \"c66f29c7-9e93-4417-bc3e-b486a0b167cc\") " Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.593726 4808 scope.go:117] "RemoveContainer" containerID="0c3523d23eea77f3a26989699d6139b4952958e86d6780f8cad3f7bce59f1a19" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.594618 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c66f29c7-9e93-4417-bc3e-b486a0b167cc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c66f29c7-9e93-4417-bc3e-b486a0b167cc" (UID: "c66f29c7-9e93-4417-bc3e-b486a0b167cc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.594919 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c66f29c7-9e93-4417-bc3e-b486a0b167cc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c66f29c7-9e93-4417-bc3e-b486a0b167cc" (UID: "c66f29c7-9e93-4417-bc3e-b486a0b167cc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.597590 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-scripts" (OuterVolumeSpecName: "scripts") pod "c66f29c7-9e93-4417-bc3e-b486a0b167cc" (UID: "c66f29c7-9e93-4417-bc3e-b486a0b167cc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.598489 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c66f29c7-9e93-4417-bc3e-b486a0b167cc-kube-api-access-wcnsx" (OuterVolumeSpecName: "kube-api-access-wcnsx") pod "c66f29c7-9e93-4417-bc3e-b486a0b167cc" (UID: "c66f29c7-9e93-4417-bc3e-b486a0b167cc"). InnerVolumeSpecName "kube-api-access-wcnsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.618158 4808 scope.go:117] "RemoveContainer" containerID="0f6479ff535d02280faff7cca75dffddfd4ae0daef90502722fe17ad3ffaa3fb" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.637387 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c66f29c7-9e93-4417-bc3e-b486a0b167cc" (UID: "c66f29c7-9e93-4417-bc3e-b486a0b167cc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.637551 4808 scope.go:117] "RemoveContainer" containerID="95e245e2a282a68d227f87f0a4202fabdd57814fc3e943d366c73a22d098be10" Oct 02 17:00:44 crc kubenswrapper[4808]: E1002 17:00:44.638545 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95e245e2a282a68d227f87f0a4202fabdd57814fc3e943d366c73a22d098be10\": container with ID starting with 95e245e2a282a68d227f87f0a4202fabdd57814fc3e943d366c73a22d098be10 not found: ID does not exist" containerID="95e245e2a282a68d227f87f0a4202fabdd57814fc3e943d366c73a22d098be10" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.638615 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95e245e2a282a68d227f87f0a4202fabdd57814fc3e943d366c73a22d098be10"} err="failed to get container status \"95e245e2a282a68d227f87f0a4202fabdd57814fc3e943d366c73a22d098be10\": rpc error: code = NotFound desc = could not find container \"95e245e2a282a68d227f87f0a4202fabdd57814fc3e943d366c73a22d098be10\": container with ID starting with 95e245e2a282a68d227f87f0a4202fabdd57814fc3e943d366c73a22d098be10 not found: ID does not exist" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.638647 4808 scope.go:117] "RemoveContainer" containerID="c097f25e6215428c150710efb5bf442f8649de49ea19162ee15380e03db6bfbd" Oct 02 17:00:44 crc kubenswrapper[4808]: E1002 17:00:44.638989 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c097f25e6215428c150710efb5bf442f8649de49ea19162ee15380e03db6bfbd\": container with ID starting with c097f25e6215428c150710efb5bf442f8649de49ea19162ee15380e03db6bfbd not found: ID does not exist" containerID="c097f25e6215428c150710efb5bf442f8649de49ea19162ee15380e03db6bfbd" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.639073 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c097f25e6215428c150710efb5bf442f8649de49ea19162ee15380e03db6bfbd"} err="failed to get container status \"c097f25e6215428c150710efb5bf442f8649de49ea19162ee15380e03db6bfbd\": rpc error: code = NotFound desc = could not find container \"c097f25e6215428c150710efb5bf442f8649de49ea19162ee15380e03db6bfbd\": container with ID starting with c097f25e6215428c150710efb5bf442f8649de49ea19162ee15380e03db6bfbd not found: ID does not exist" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.639100 4808 scope.go:117] "RemoveContainer" containerID="0c3523d23eea77f3a26989699d6139b4952958e86d6780f8cad3f7bce59f1a19" Oct 02 17:00:44 crc kubenswrapper[4808]: E1002 17:00:44.639365 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c3523d23eea77f3a26989699d6139b4952958e86d6780f8cad3f7bce59f1a19\": container with ID starting with 0c3523d23eea77f3a26989699d6139b4952958e86d6780f8cad3f7bce59f1a19 not found: ID does not exist" containerID="0c3523d23eea77f3a26989699d6139b4952958e86d6780f8cad3f7bce59f1a19" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.639397 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c3523d23eea77f3a26989699d6139b4952958e86d6780f8cad3f7bce59f1a19"} err="failed to get container status \"0c3523d23eea77f3a26989699d6139b4952958e86d6780f8cad3f7bce59f1a19\": rpc error: code = NotFound desc = could not find container \"0c3523d23eea77f3a26989699d6139b4952958e86d6780f8cad3f7bce59f1a19\": container with ID starting with 0c3523d23eea77f3a26989699d6139b4952958e86d6780f8cad3f7bce59f1a19 not found: ID does not exist" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.639415 4808 scope.go:117] "RemoveContainer" containerID="0f6479ff535d02280faff7cca75dffddfd4ae0daef90502722fe17ad3ffaa3fb" Oct 02 17:00:44 crc kubenswrapper[4808]: E1002 17:00:44.639662 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f6479ff535d02280faff7cca75dffddfd4ae0daef90502722fe17ad3ffaa3fb\": container with ID starting with 0f6479ff535d02280faff7cca75dffddfd4ae0daef90502722fe17ad3ffaa3fb not found: ID does not exist" containerID="0f6479ff535d02280faff7cca75dffddfd4ae0daef90502722fe17ad3ffaa3fb" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.639691 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f6479ff535d02280faff7cca75dffddfd4ae0daef90502722fe17ad3ffaa3fb"} err="failed to get container status \"0f6479ff535d02280faff7cca75dffddfd4ae0daef90502722fe17ad3ffaa3fb\": rpc error: code = NotFound desc = could not find container \"0f6479ff535d02280faff7cca75dffddfd4ae0daef90502722fe17ad3ffaa3fb\": container with ID starting with 0f6479ff535d02280faff7cca75dffddfd4ae0daef90502722fe17ad3ffaa3fb not found: ID does not exist" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.653797 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "c66f29c7-9e93-4417-bc3e-b486a0b167cc" (UID: "c66f29c7-9e93-4417-bc3e-b486a0b167cc"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.684965 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c66f29c7-9e93-4417-bc3e-b486a0b167cc" (UID: "c66f29c7-9e93-4417-bc3e-b486a0b167cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.697224 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.697272 4808 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c66f29c7-9e93-4417-bc3e-b486a0b167cc-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.697286 4808 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.697298 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.697310 4808 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.697323 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcnsx\" (UniqueName: \"kubernetes.io/projected/c66f29c7-9e93-4417-bc3e-b486a0b167cc-kube-api-access-wcnsx\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.697334 4808 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c66f29c7-9e93-4417-bc3e-b486a0b167cc-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.699849 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-config-data" (OuterVolumeSpecName: "config-data") pod "c66f29c7-9e93-4417-bc3e-b486a0b167cc" (UID: "c66f29c7-9e93-4417-bc3e-b486a0b167cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.798664 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c66f29c7-9e93-4417-bc3e-b486a0b167cc-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.906708 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.917665 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.938114 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 17:00:44 crc kubenswrapper[4808]: E1002 17:00:44.938518 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" containerName="ceilometer-notification-agent" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.938536 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" containerName="ceilometer-notification-agent" Oct 02 17:00:44 crc kubenswrapper[4808]: E1002 17:00:44.938550 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" containerName="sg-core" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.938557 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" containerName="sg-core" Oct 02 17:00:44 crc kubenswrapper[4808]: E1002 17:00:44.938586 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" containerName="ceilometer-central-agent" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.938593 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" containerName="ceilometer-central-agent" Oct 02 17:00:44 crc kubenswrapper[4808]: E1002 17:00:44.938607 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" containerName="proxy-httpd" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.938613 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" containerName="proxy-httpd" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.938771 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" containerName="sg-core" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.938785 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" containerName="ceilometer-central-agent" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.938795 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" containerName="proxy-httpd" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.938803 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" containerName="ceilometer-notification-agent" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.940713 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.947755 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.948000 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.948066 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 17:00:44 crc kubenswrapper[4808]: I1002 17:00:44.979634 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.109663 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33df7922-0de2-4897-9cfa-599fadc6229f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.109705 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/33df7922-0de2-4897-9cfa-599fadc6229f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.109740 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33df7922-0de2-4897-9cfa-599fadc6229f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.109760 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33df7922-0de2-4897-9cfa-599fadc6229f-run-httpd\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.109803 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33df7922-0de2-4897-9cfa-599fadc6229f-scripts\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.109830 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33df7922-0de2-4897-9cfa-599fadc6229f-config-data\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.109952 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33df7922-0de2-4897-9cfa-599fadc6229f-log-httpd\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.110011 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8fb7\" (UniqueName: \"kubernetes.io/projected/33df7922-0de2-4897-9cfa-599fadc6229f-kube-api-access-d8fb7\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.212446 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33df7922-0de2-4897-9cfa-599fadc6229f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.212572 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/33df7922-0de2-4897-9cfa-599fadc6229f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.212736 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33df7922-0de2-4897-9cfa-599fadc6229f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.212827 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33df7922-0de2-4897-9cfa-599fadc6229f-run-httpd\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.212955 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33df7922-0de2-4897-9cfa-599fadc6229f-scripts\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.213040 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33df7922-0de2-4897-9cfa-599fadc6229f-config-data\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.213119 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33df7922-0de2-4897-9cfa-599fadc6229f-log-httpd\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.213185 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8fb7\" (UniqueName: \"kubernetes.io/projected/33df7922-0de2-4897-9cfa-599fadc6229f-kube-api-access-d8fb7\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.213553 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33df7922-0de2-4897-9cfa-599fadc6229f-run-httpd\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.213976 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33df7922-0de2-4897-9cfa-599fadc6229f-log-httpd\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.218910 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33df7922-0de2-4897-9cfa-599fadc6229f-scripts\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.218969 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33df7922-0de2-4897-9cfa-599fadc6229f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.219305 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33df7922-0de2-4897-9cfa-599fadc6229f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.220431 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33df7922-0de2-4897-9cfa-599fadc6229f-config-data\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.231765 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/33df7922-0de2-4897-9cfa-599fadc6229f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.237384 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8fb7\" (UniqueName: \"kubernetes.io/projected/33df7922-0de2-4897-9cfa-599fadc6229f-kube-api-access-d8fb7\") pod \"ceilometer-0\" (UID: \"33df7922-0de2-4897-9cfa-599fadc6229f\") " pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.260989 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.408297 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c66f29c7-9e93-4417-bc3e-b486a0b167cc" path="/var/lib/kubelet/pods/c66f29c7-9e93-4417-bc3e-b486a0b167cc/volumes" Oct 02 17:00:45 crc kubenswrapper[4808]: W1002 17:00:45.820269 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33df7922_0de2_4897_9cfa_599fadc6229f.slice/crio-57b36d39af95611db529de69d665bbc9759e95c9bda71302ec5676e1ac966ff2 WatchSource:0}: Error finding container 57b36d39af95611db529de69d665bbc9759e95c9bda71302ec5676e1ac966ff2: Status 404 returned error can't find the container with id 57b36d39af95611db529de69d665bbc9759e95c9bda71302ec5676e1ac966ff2 Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.823366 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.824950 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 17:00:45 crc kubenswrapper[4808]: I1002 17:00:45.866868 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 17:00:46 crc kubenswrapper[4808]: I1002 17:00:46.560308 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33df7922-0de2-4897-9cfa-599fadc6229f","Type":"ContainerStarted","Data":"57b36d39af95611db529de69d665bbc9759e95c9bda71302ec5676e1ac966ff2"} Oct 02 17:00:47 crc kubenswrapper[4808]: I1002 17:00:47.569182 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33df7922-0de2-4897-9cfa-599fadc6229f","Type":"ContainerStarted","Data":"b5223bb0baf68ed1e97d01c09cc4b751cdeb7362801e3844e1a15d03e82d527c"} Oct 02 17:00:47 crc kubenswrapper[4808]: I1002 17:00:47.569464 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33df7922-0de2-4897-9cfa-599fadc6229f","Type":"ContainerStarted","Data":"7c8146cb5c9664098f7b0cca677f0bed5b720e3408b1980a7f7138d0f397b53a"} Oct 02 17:00:48 crc kubenswrapper[4808]: I1002 17:00:48.224808 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 17:00:48 crc kubenswrapper[4808]: I1002 17:00:48.225091 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 17:00:48 crc kubenswrapper[4808]: I1002 17:00:48.580338 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33df7922-0de2-4897-9cfa-599fadc6229f","Type":"ContainerStarted","Data":"fa0ba59f990dba9a37957a728206e0f05b7d8c9990dbf929dd118581e7e69975"} Oct 02 17:00:48 crc kubenswrapper[4808]: I1002 17:00:48.739681 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 17:00:48 crc kubenswrapper[4808]: I1002 17:00:48.739734 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 17:00:50 crc kubenswrapper[4808]: I1002 17:00:50.223180 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 17:00:50 crc kubenswrapper[4808]: I1002 17:00:50.223454 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 17:00:50 crc kubenswrapper[4808]: I1002 17:00:50.606425 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33df7922-0de2-4897-9cfa-599fadc6229f","Type":"ContainerStarted","Data":"33cdbc994fc86b666d00ffc250e5f81424cb696c7ccb4e5d65ae409b0b14c4b5"} Oct 02 17:00:50 crc kubenswrapper[4808]: I1002 17:00:50.608437 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 17:00:50 crc kubenswrapper[4808]: I1002 17:00:50.652143 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.826880746 podStartE2EDuration="6.652122912s" podCreationTimestamp="2025-10-02 17:00:44 +0000 UTC" firstStartedPulling="2025-10-02 17:00:45.824770277 +0000 UTC m=+1213.150299267" lastFinishedPulling="2025-10-02 17:00:49.650012403 +0000 UTC m=+1216.975541433" observedRunningTime="2025-10-02 17:00:50.639983317 +0000 UTC m=+1217.965512357" watchObservedRunningTime="2025-10-02 17:00:50.652122912 +0000 UTC m=+1217.977651912" Oct 02 17:00:50 crc kubenswrapper[4808]: I1002 17:00:50.866553 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 17:00:50 crc kubenswrapper[4808]: I1002 17:00:50.910841 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 17:00:51 crc kubenswrapper[4808]: I1002 17:00:51.242383 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fa5606b5-d0de-4403-b815-41d3216754d5" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.187:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 17:00:51 crc kubenswrapper[4808]: I1002 17:00:51.242454 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fa5606b5-d0de-4403-b815-41d3216754d5" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.187:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 17:00:51 crc kubenswrapper[4808]: I1002 17:00:51.641841 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 17:00:53 crc kubenswrapper[4808]: I1002 17:00:53.224880 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 17:00:53 crc kubenswrapper[4808]: I1002 17:00:53.225332 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 17:00:54 crc kubenswrapper[4808]: I1002 17:00:54.243450 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c0748c83-a144-4c4c-9980-41c6592edf0f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 17:00:54 crc kubenswrapper[4808]: I1002 17:00:54.243461 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c0748c83-a144-4c4c-9980-41c6592edf0f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.160425 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29323741-x5mvf"] Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.164609 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323741-x5mvf" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.184127 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323741-x5mvf"] Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.231131 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.231933 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.238162 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.238516 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.281814 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/765ddaf0-84a9-4ec2-b07e-683ebe340353-fernet-keys\") pod \"keystone-cron-29323741-x5mvf\" (UID: \"765ddaf0-84a9-4ec2-b07e-683ebe340353\") " pod="openstack/keystone-cron-29323741-x5mvf" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.282110 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f744g\" (UniqueName: \"kubernetes.io/projected/765ddaf0-84a9-4ec2-b07e-683ebe340353-kube-api-access-f744g\") pod \"keystone-cron-29323741-x5mvf\" (UID: \"765ddaf0-84a9-4ec2-b07e-683ebe340353\") " pod="openstack/keystone-cron-29323741-x5mvf" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.282161 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/765ddaf0-84a9-4ec2-b07e-683ebe340353-config-data\") pod \"keystone-cron-29323741-x5mvf\" (UID: \"765ddaf0-84a9-4ec2-b07e-683ebe340353\") " pod="openstack/keystone-cron-29323741-x5mvf" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.282316 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/765ddaf0-84a9-4ec2-b07e-683ebe340353-combined-ca-bundle\") pod \"keystone-cron-29323741-x5mvf\" (UID: \"765ddaf0-84a9-4ec2-b07e-683ebe340353\") " pod="openstack/keystone-cron-29323741-x5mvf" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.384558 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f744g\" (UniqueName: \"kubernetes.io/projected/765ddaf0-84a9-4ec2-b07e-683ebe340353-kube-api-access-f744g\") pod \"keystone-cron-29323741-x5mvf\" (UID: \"765ddaf0-84a9-4ec2-b07e-683ebe340353\") " pod="openstack/keystone-cron-29323741-x5mvf" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.384664 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/765ddaf0-84a9-4ec2-b07e-683ebe340353-config-data\") pod \"keystone-cron-29323741-x5mvf\" (UID: \"765ddaf0-84a9-4ec2-b07e-683ebe340353\") " pod="openstack/keystone-cron-29323741-x5mvf" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.384907 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/765ddaf0-84a9-4ec2-b07e-683ebe340353-combined-ca-bundle\") pod \"keystone-cron-29323741-x5mvf\" (UID: \"765ddaf0-84a9-4ec2-b07e-683ebe340353\") " pod="openstack/keystone-cron-29323741-x5mvf" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.385012 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/765ddaf0-84a9-4ec2-b07e-683ebe340353-fernet-keys\") pod \"keystone-cron-29323741-x5mvf\" (UID: \"765ddaf0-84a9-4ec2-b07e-683ebe340353\") " pod="openstack/keystone-cron-29323741-x5mvf" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.392116 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/765ddaf0-84a9-4ec2-b07e-683ebe340353-config-data\") pod \"keystone-cron-29323741-x5mvf\" (UID: \"765ddaf0-84a9-4ec2-b07e-683ebe340353\") " pod="openstack/keystone-cron-29323741-x5mvf" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.392966 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/765ddaf0-84a9-4ec2-b07e-683ebe340353-fernet-keys\") pod \"keystone-cron-29323741-x5mvf\" (UID: \"765ddaf0-84a9-4ec2-b07e-683ebe340353\") " pod="openstack/keystone-cron-29323741-x5mvf" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.394105 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/765ddaf0-84a9-4ec2-b07e-683ebe340353-combined-ca-bundle\") pod \"keystone-cron-29323741-x5mvf\" (UID: \"765ddaf0-84a9-4ec2-b07e-683ebe340353\") " pod="openstack/keystone-cron-29323741-x5mvf" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.411873 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f744g\" (UniqueName: \"kubernetes.io/projected/765ddaf0-84a9-4ec2-b07e-683ebe340353-kube-api-access-f744g\") pod \"keystone-cron-29323741-x5mvf\" (UID: \"765ddaf0-84a9-4ec2-b07e-683ebe340353\") " pod="openstack/keystone-cron-29323741-x5mvf" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.485903 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323741-x5mvf" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.722607 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.731096 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 17:01:00 crc kubenswrapper[4808]: I1002 17:01:00.963113 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323741-x5mvf"] Oct 02 17:01:00 crc kubenswrapper[4808]: W1002 17:01:00.972781 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod765ddaf0_84a9_4ec2_b07e_683ebe340353.slice/crio-0ca802771882e7be656f79ee70e2e9e91b6ed3846b0920d1a527866698d4aae1 WatchSource:0}: Error finding container 0ca802771882e7be656f79ee70e2e9e91b6ed3846b0920d1a527866698d4aae1: Status 404 returned error can't find the container with id 0ca802771882e7be656f79ee70e2e9e91b6ed3846b0920d1a527866698d4aae1 Oct 02 17:01:01 crc kubenswrapper[4808]: I1002 17:01:01.734908 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323741-x5mvf" event={"ID":"765ddaf0-84a9-4ec2-b07e-683ebe340353","Type":"ContainerStarted","Data":"97cb2808a316a0b7eb12116058fcf65301e1efb5860dad6388951dcb710b527a"} Oct 02 17:01:01 crc kubenswrapper[4808]: I1002 17:01:01.735484 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323741-x5mvf" event={"ID":"765ddaf0-84a9-4ec2-b07e-683ebe340353","Type":"ContainerStarted","Data":"0ca802771882e7be656f79ee70e2e9e91b6ed3846b0920d1a527866698d4aae1"} Oct 02 17:01:01 crc kubenswrapper[4808]: I1002 17:01:01.772583 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29323741-x5mvf" podStartSLOduration=1.772550879 podStartE2EDuration="1.772550879s" podCreationTimestamp="2025-10-02 17:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 17:01:01.758338348 +0000 UTC m=+1229.083867428" watchObservedRunningTime="2025-10-02 17:01:01.772550879 +0000 UTC m=+1229.098079909" Oct 02 17:01:03 crc kubenswrapper[4808]: I1002 17:01:03.230344 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 17:01:03 crc kubenswrapper[4808]: I1002 17:01:03.233029 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 17:01:03 crc kubenswrapper[4808]: I1002 17:01:03.237578 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 17:01:03 crc kubenswrapper[4808]: I1002 17:01:03.757685 4808 generic.go:334] "Generic (PLEG): container finished" podID="765ddaf0-84a9-4ec2-b07e-683ebe340353" containerID="97cb2808a316a0b7eb12116058fcf65301e1efb5860dad6388951dcb710b527a" exitCode=0 Oct 02 17:01:03 crc kubenswrapper[4808]: I1002 17:01:03.757752 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323741-x5mvf" event={"ID":"765ddaf0-84a9-4ec2-b07e-683ebe340353","Type":"ContainerDied","Data":"97cb2808a316a0b7eb12116058fcf65301e1efb5860dad6388951dcb710b527a"} Oct 02 17:01:03 crc kubenswrapper[4808]: I1002 17:01:03.763631 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 17:01:05 crc kubenswrapper[4808]: I1002 17:01:05.161932 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323741-x5mvf" Oct 02 17:01:05 crc kubenswrapper[4808]: I1002 17:01:05.281299 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/765ddaf0-84a9-4ec2-b07e-683ebe340353-combined-ca-bundle\") pod \"765ddaf0-84a9-4ec2-b07e-683ebe340353\" (UID: \"765ddaf0-84a9-4ec2-b07e-683ebe340353\") " Oct 02 17:01:05 crc kubenswrapper[4808]: I1002 17:01:05.281371 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/765ddaf0-84a9-4ec2-b07e-683ebe340353-fernet-keys\") pod \"765ddaf0-84a9-4ec2-b07e-683ebe340353\" (UID: \"765ddaf0-84a9-4ec2-b07e-683ebe340353\") " Oct 02 17:01:05 crc kubenswrapper[4808]: I1002 17:01:05.281400 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/765ddaf0-84a9-4ec2-b07e-683ebe340353-config-data\") pod \"765ddaf0-84a9-4ec2-b07e-683ebe340353\" (UID: \"765ddaf0-84a9-4ec2-b07e-683ebe340353\") " Oct 02 17:01:05 crc kubenswrapper[4808]: I1002 17:01:05.281513 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f744g\" (UniqueName: \"kubernetes.io/projected/765ddaf0-84a9-4ec2-b07e-683ebe340353-kube-api-access-f744g\") pod \"765ddaf0-84a9-4ec2-b07e-683ebe340353\" (UID: \"765ddaf0-84a9-4ec2-b07e-683ebe340353\") " Oct 02 17:01:05 crc kubenswrapper[4808]: I1002 17:01:05.288292 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/765ddaf0-84a9-4ec2-b07e-683ebe340353-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "765ddaf0-84a9-4ec2-b07e-683ebe340353" (UID: "765ddaf0-84a9-4ec2-b07e-683ebe340353"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:01:05 crc kubenswrapper[4808]: I1002 17:01:05.292777 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/765ddaf0-84a9-4ec2-b07e-683ebe340353-kube-api-access-f744g" (OuterVolumeSpecName: "kube-api-access-f744g") pod "765ddaf0-84a9-4ec2-b07e-683ebe340353" (UID: "765ddaf0-84a9-4ec2-b07e-683ebe340353"). InnerVolumeSpecName "kube-api-access-f744g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:01:05 crc kubenswrapper[4808]: I1002 17:01:05.335788 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/765ddaf0-84a9-4ec2-b07e-683ebe340353-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "765ddaf0-84a9-4ec2-b07e-683ebe340353" (UID: "765ddaf0-84a9-4ec2-b07e-683ebe340353"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:01:05 crc kubenswrapper[4808]: I1002 17:01:05.365127 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/765ddaf0-84a9-4ec2-b07e-683ebe340353-config-data" (OuterVolumeSpecName: "config-data") pod "765ddaf0-84a9-4ec2-b07e-683ebe340353" (UID: "765ddaf0-84a9-4ec2-b07e-683ebe340353"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:01:05 crc kubenswrapper[4808]: I1002 17:01:05.383652 4808 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/765ddaf0-84a9-4ec2-b07e-683ebe340353-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:05 crc kubenswrapper[4808]: I1002 17:01:05.383693 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/765ddaf0-84a9-4ec2-b07e-683ebe340353-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:05 crc kubenswrapper[4808]: I1002 17:01:05.383714 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f744g\" (UniqueName: \"kubernetes.io/projected/765ddaf0-84a9-4ec2-b07e-683ebe340353-kube-api-access-f744g\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:05 crc kubenswrapper[4808]: I1002 17:01:05.383735 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/765ddaf0-84a9-4ec2-b07e-683ebe340353-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:05 crc kubenswrapper[4808]: I1002 17:01:05.779406 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323741-x5mvf" event={"ID":"765ddaf0-84a9-4ec2-b07e-683ebe340353","Type":"ContainerDied","Data":"0ca802771882e7be656f79ee70e2e9e91b6ed3846b0920d1a527866698d4aae1"} Oct 02 17:01:05 crc kubenswrapper[4808]: I1002 17:01:05.779857 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ca802771882e7be656f79ee70e2e9e91b6ed3846b0920d1a527866698d4aae1" Oct 02 17:01:05 crc kubenswrapper[4808]: I1002 17:01:05.779581 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323741-x5mvf" Oct 02 17:01:15 crc kubenswrapper[4808]: I1002 17:01:15.272140 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 17:01:18 crc kubenswrapper[4808]: I1002 17:01:18.739667 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 17:01:18 crc kubenswrapper[4808]: I1002 17:01:18.740124 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 17:01:25 crc kubenswrapper[4808]: I1002 17:01:25.048265 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 17:01:26 crc kubenswrapper[4808]: I1002 17:01:26.514128 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 17:01:29 crc kubenswrapper[4808]: I1002 17:01:29.057434 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="f71818d6-1ebf-4838-b871-8873c107b325" containerName="rabbitmq" containerID="cri-o://c4f8b08ae8e573d59cb3204c0cdb27282c5471beafd2a27b8cc8c7464f5d768a" gracePeriod=604796 Oct 02 17:01:30 crc kubenswrapper[4808]: I1002 17:01:30.649432 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="b409b3d8-0ed6-49e9-a1ea-be6689e607ed" containerName="rabbitmq" containerID="cri-o://516b218bf929f18eb0453704db63a9a632bf10bea5cd9e6b24030634967a64fa" gracePeriod=604796 Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.135622 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="f71818d6-1ebf-4838-b871-8873c107b325" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.436678 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="b409b3d8-0ed6-49e9-a1ea-be6689e607ed" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.617474 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.784797 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f71818d6-1ebf-4838-b871-8873c107b325-pod-info\") pod \"f71818d6-1ebf-4838-b871-8873c107b325\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.784866 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-plugins\") pod \"f71818d6-1ebf-4838-b871-8873c107b325\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.784902 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-confd\") pod \"f71818d6-1ebf-4838-b871-8873c107b325\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.784932 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-tls\") pod \"f71818d6-1ebf-4838-b871-8873c107b325\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.784982 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tvxc\" (UniqueName: \"kubernetes.io/projected/f71818d6-1ebf-4838-b871-8873c107b325-kube-api-access-5tvxc\") pod \"f71818d6-1ebf-4838-b871-8873c107b325\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.785018 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f71818d6-1ebf-4838-b871-8873c107b325-plugins-conf\") pod \"f71818d6-1ebf-4838-b871-8873c107b325\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.785038 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f71818d6-1ebf-4838-b871-8873c107b325-server-conf\") pod \"f71818d6-1ebf-4838-b871-8873c107b325\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.785122 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f71818d6-1ebf-4838-b871-8873c107b325-config-data\") pod \"f71818d6-1ebf-4838-b871-8873c107b325\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.785146 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-erlang-cookie\") pod \"f71818d6-1ebf-4838-b871-8873c107b325\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.785178 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f71818d6-1ebf-4838-b871-8873c107b325-erlang-cookie-secret\") pod \"f71818d6-1ebf-4838-b871-8873c107b325\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.785204 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"f71818d6-1ebf-4838-b871-8873c107b325\" (UID: \"f71818d6-1ebf-4838-b871-8873c107b325\") " Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.785653 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "f71818d6-1ebf-4838-b871-8873c107b325" (UID: "f71818d6-1ebf-4838-b871-8873c107b325"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.785657 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f71818d6-1ebf-4838-b871-8873c107b325-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "f71818d6-1ebf-4838-b871-8873c107b325" (UID: "f71818d6-1ebf-4838-b871-8873c107b325"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.785824 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "f71818d6-1ebf-4838-b871-8873c107b325" (UID: "f71818d6-1ebf-4838-b871-8873c107b325"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.786034 4808 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.786046 4808 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f71818d6-1ebf-4838-b871-8873c107b325-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.786055 4808 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.792007 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "persistence") pod "f71818d6-1ebf-4838-b871-8873c107b325" (UID: "f71818d6-1ebf-4838-b871-8873c107b325"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.792176 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f71818d6-1ebf-4838-b871-8873c107b325-kube-api-access-5tvxc" (OuterVolumeSpecName: "kube-api-access-5tvxc") pod "f71818d6-1ebf-4838-b871-8873c107b325" (UID: "f71818d6-1ebf-4838-b871-8873c107b325"). InnerVolumeSpecName "kube-api-access-5tvxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.792552 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f71818d6-1ebf-4838-b871-8873c107b325-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "f71818d6-1ebf-4838-b871-8873c107b325" (UID: "f71818d6-1ebf-4838-b871-8873c107b325"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.811387 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/f71818d6-1ebf-4838-b871-8873c107b325-pod-info" (OuterVolumeSpecName: "pod-info") pod "f71818d6-1ebf-4838-b871-8873c107b325" (UID: "f71818d6-1ebf-4838-b871-8873c107b325"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.826079 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "f71818d6-1ebf-4838-b871-8873c107b325" (UID: "f71818d6-1ebf-4838-b871-8873c107b325"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.848056 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f71818d6-1ebf-4838-b871-8873c107b325-config-data" (OuterVolumeSpecName: "config-data") pod "f71818d6-1ebf-4838-b871-8873c107b325" (UID: "f71818d6-1ebf-4838-b871-8873c107b325"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.858172 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f71818d6-1ebf-4838-b871-8873c107b325-server-conf" (OuterVolumeSpecName: "server-conf") pod "f71818d6-1ebf-4838-b871-8873c107b325" (UID: "f71818d6-1ebf-4838-b871-8873c107b325"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.887904 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tvxc\" (UniqueName: \"kubernetes.io/projected/f71818d6-1ebf-4838-b871-8873c107b325-kube-api-access-5tvxc\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.887935 4808 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f71818d6-1ebf-4838-b871-8873c107b325-server-conf\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.887947 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f71818d6-1ebf-4838-b871-8873c107b325-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.887956 4808 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f71818d6-1ebf-4838-b871-8873c107b325-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.887983 4808 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.887991 4808 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f71818d6-1ebf-4838-b871-8873c107b325-pod-info\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.887999 4808 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.907786 4808 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.918753 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "f71818d6-1ebf-4838-b871-8873c107b325" (UID: "f71818d6-1ebf-4838-b871-8873c107b325"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.989502 4808 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:35 crc kubenswrapper[4808]: I1002 17:01:35.990332 4808 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f71818d6-1ebf-4838-b871-8873c107b325-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.128456 4808 generic.go:334] "Generic (PLEG): container finished" podID="f71818d6-1ebf-4838-b871-8873c107b325" containerID="c4f8b08ae8e573d59cb3204c0cdb27282c5471beafd2a27b8cc8c7464f5d768a" exitCode=0 Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.128512 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f71818d6-1ebf-4838-b871-8873c107b325","Type":"ContainerDied","Data":"c4f8b08ae8e573d59cb3204c0cdb27282c5471beafd2a27b8cc8c7464f5d768a"} Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.128752 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f71818d6-1ebf-4838-b871-8873c107b325","Type":"ContainerDied","Data":"3fa89b1cc3dfc663ebc7dec8bcfefa85feb57fd1577efc4daeff8f04db62989f"} Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.128575 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.128820 4808 scope.go:117] "RemoveContainer" containerID="c4f8b08ae8e573d59cb3204c0cdb27282c5471beafd2a27b8cc8c7464f5d768a" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.151618 4808 scope.go:117] "RemoveContainer" containerID="3733dae6ab12fcdb987673b7dd9fdcbf7a55150a8de5aa8be5de065407e9afcf" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.178683 4808 scope.go:117] "RemoveContainer" containerID="c4f8b08ae8e573d59cb3204c0cdb27282c5471beafd2a27b8cc8c7464f5d768a" Oct 02 17:01:36 crc kubenswrapper[4808]: E1002 17:01:36.179026 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4f8b08ae8e573d59cb3204c0cdb27282c5471beafd2a27b8cc8c7464f5d768a\": container with ID starting with c4f8b08ae8e573d59cb3204c0cdb27282c5471beafd2a27b8cc8c7464f5d768a not found: ID does not exist" containerID="c4f8b08ae8e573d59cb3204c0cdb27282c5471beafd2a27b8cc8c7464f5d768a" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.179063 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4f8b08ae8e573d59cb3204c0cdb27282c5471beafd2a27b8cc8c7464f5d768a"} err="failed to get container status \"c4f8b08ae8e573d59cb3204c0cdb27282c5471beafd2a27b8cc8c7464f5d768a\": rpc error: code = NotFound desc = could not find container \"c4f8b08ae8e573d59cb3204c0cdb27282c5471beafd2a27b8cc8c7464f5d768a\": container with ID starting with c4f8b08ae8e573d59cb3204c0cdb27282c5471beafd2a27b8cc8c7464f5d768a not found: ID does not exist" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.179086 4808 scope.go:117] "RemoveContainer" containerID="3733dae6ab12fcdb987673b7dd9fdcbf7a55150a8de5aa8be5de065407e9afcf" Oct 02 17:01:36 crc kubenswrapper[4808]: E1002 17:01:36.180583 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3733dae6ab12fcdb987673b7dd9fdcbf7a55150a8de5aa8be5de065407e9afcf\": container with ID starting with 3733dae6ab12fcdb987673b7dd9fdcbf7a55150a8de5aa8be5de065407e9afcf not found: ID does not exist" containerID="3733dae6ab12fcdb987673b7dd9fdcbf7a55150a8de5aa8be5de065407e9afcf" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.180621 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3733dae6ab12fcdb987673b7dd9fdcbf7a55150a8de5aa8be5de065407e9afcf"} err="failed to get container status \"3733dae6ab12fcdb987673b7dd9fdcbf7a55150a8de5aa8be5de065407e9afcf\": rpc error: code = NotFound desc = could not find container \"3733dae6ab12fcdb987673b7dd9fdcbf7a55150a8de5aa8be5de065407e9afcf\": container with ID starting with 3733dae6ab12fcdb987673b7dd9fdcbf7a55150a8de5aa8be5de065407e9afcf not found: ID does not exist" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.182099 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.203755 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.216985 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 17:01:36 crc kubenswrapper[4808]: E1002 17:01:36.217337 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f71818d6-1ebf-4838-b871-8873c107b325" containerName="setup-container" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.217349 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f71818d6-1ebf-4838-b871-8873c107b325" containerName="setup-container" Oct 02 17:01:36 crc kubenswrapper[4808]: E1002 17:01:36.217361 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="765ddaf0-84a9-4ec2-b07e-683ebe340353" containerName="keystone-cron" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.217367 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="765ddaf0-84a9-4ec2-b07e-683ebe340353" containerName="keystone-cron" Oct 02 17:01:36 crc kubenswrapper[4808]: E1002 17:01:36.217384 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f71818d6-1ebf-4838-b871-8873c107b325" containerName="rabbitmq" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.217390 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f71818d6-1ebf-4838-b871-8873c107b325" containerName="rabbitmq" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.217536 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="765ddaf0-84a9-4ec2-b07e-683ebe340353" containerName="keystone-cron" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.217557 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f71818d6-1ebf-4838-b871-8873c107b325" containerName="rabbitmq" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.218391 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.227474 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.227528 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.227701 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.227713 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.227823 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.227828 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-xzjch" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.227971 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.238253 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.398721 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.398800 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-server-conf\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.398898 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.398935 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-config-data\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.398982 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.399023 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-pod-info\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.399053 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzhvd\" (UniqueName: \"kubernetes.io/projected/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-kube-api-access-pzhvd\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.399088 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.399143 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.399171 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.399314 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.501493 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.501562 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-server-conf\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.501616 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.501633 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-config-data\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.501650 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.501677 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-pod-info\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.501691 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.501708 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzhvd\" (UniqueName: \"kubernetes.io/projected/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-kube-api-access-pzhvd\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.501742 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.501757 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.501778 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.501781 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.502450 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.503313 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.503606 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.503924 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-config-data\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.504533 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-server-conf\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.507323 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.507584 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.509285 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-pod-info\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.515774 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.530325 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzhvd\" (UniqueName: \"kubernetes.io/projected/99aebb2e-33d8-4657-9ed2-cf23aeb8a281-kube-api-access-pzhvd\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.557995 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"99aebb2e-33d8-4657-9ed2-cf23aeb8a281\") " pod="openstack/rabbitmq-server-0" Oct 02 17:01:36 crc kubenswrapper[4808]: I1002 17:01:36.843445 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.138089 4808 generic.go:334] "Generic (PLEG): container finished" podID="b409b3d8-0ed6-49e9-a1ea-be6689e607ed" containerID="516b218bf929f18eb0453704db63a9a632bf10bea5cd9e6b24030634967a64fa" exitCode=0 Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.138352 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b409b3d8-0ed6-49e9-a1ea-be6689e607ed","Type":"ContainerDied","Data":"516b218bf929f18eb0453704db63a9a632bf10bea5cd9e6b24030634967a64fa"} Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.277988 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.407984 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f71818d6-1ebf-4838-b871-8873c107b325" path="/var/lib/kubelet/pods/f71818d6-1ebf-4838-b871-8873c107b325/volumes" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.418105 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.418189 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-pod-info\") pod \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.418225 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-erlang-cookie\") pod \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.418278 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-config-data\") pod \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.418326 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-plugins\") pod \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.418352 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-confd\") pod \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.418445 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-server-conf\") pod \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.418478 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8669v\" (UniqueName: \"kubernetes.io/projected/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-kube-api-access-8669v\") pod \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.418522 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-plugins-conf\") pod \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.418564 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-erlang-cookie-secret\") pod \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.419156 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-tls\") pod \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\" (UID: \"b409b3d8-0ed6-49e9-a1ea-be6689e607ed\") " Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.419219 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "b409b3d8-0ed6-49e9-a1ea-be6689e607ed" (UID: "b409b3d8-0ed6-49e9-a1ea-be6689e607ed"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.419267 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "b409b3d8-0ed6-49e9-a1ea-be6689e607ed" (UID: "b409b3d8-0ed6-49e9-a1ea-be6689e607ed"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.419525 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "b409b3d8-0ed6-49e9-a1ea-be6689e607ed" (UID: "b409b3d8-0ed6-49e9-a1ea-be6689e607ed"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.419951 4808 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.419974 4808 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.419989 4808 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.425893 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "b409b3d8-0ed6-49e9-a1ea-be6689e607ed" (UID: "b409b3d8-0ed6-49e9-a1ea-be6689e607ed"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.425934 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "b409b3d8-0ed6-49e9-a1ea-be6689e607ed" (UID: "b409b3d8-0ed6-49e9-a1ea-be6689e607ed"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.426035 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-pod-info" (OuterVolumeSpecName: "pod-info") pod "b409b3d8-0ed6-49e9-a1ea-be6689e607ed" (UID: "b409b3d8-0ed6-49e9-a1ea-be6689e607ed"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.427975 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-kube-api-access-8669v" (OuterVolumeSpecName: "kube-api-access-8669v") pod "b409b3d8-0ed6-49e9-a1ea-be6689e607ed" (UID: "b409b3d8-0ed6-49e9-a1ea-be6689e607ed"). InnerVolumeSpecName "kube-api-access-8669v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.428903 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "b409b3d8-0ed6-49e9-a1ea-be6689e607ed" (UID: "b409b3d8-0ed6-49e9-a1ea-be6689e607ed"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.445317 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-config-data" (OuterVolumeSpecName: "config-data") pod "b409b3d8-0ed6-49e9-a1ea-be6689e607ed" (UID: "b409b3d8-0ed6-49e9-a1ea-be6689e607ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 17:01:37 crc kubenswrapper[4808]: W1002 17:01:37.477751 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99aebb2e_33d8_4657_9ed2_cf23aeb8a281.slice/crio-9ceabf0844087365c08ba52bc044c8e5bb247b8d65fb479bb3cd3d98e46721f8 WatchSource:0}: Error finding container 9ceabf0844087365c08ba52bc044c8e5bb247b8d65fb479bb3cd3d98e46721f8: Status 404 returned error can't find the container with id 9ceabf0844087365c08ba52bc044c8e5bb247b8d65fb479bb3cd3d98e46721f8 Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.478010 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.484400 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-server-conf" (OuterVolumeSpecName: "server-conf") pod "b409b3d8-0ed6-49e9-a1ea-be6689e607ed" (UID: "b409b3d8-0ed6-49e9-a1ea-be6689e607ed"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.532334 4808 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.532654 4808 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.532681 4808 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.532693 4808 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-pod-info\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.532705 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.532716 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8669v\" (UniqueName: \"kubernetes.io/projected/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-kube-api-access-8669v\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.532729 4808 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-server-conf\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.537957 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "b409b3d8-0ed6-49e9-a1ea-be6689e607ed" (UID: "b409b3d8-0ed6-49e9-a1ea-be6689e607ed"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.553888 4808 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.634297 4808 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:37 crc kubenswrapper[4808]: I1002 17:01:37.634329 4808 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b409b3d8-0ed6-49e9-a1ea-be6689e607ed-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.157648 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.158593 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b409b3d8-0ed6-49e9-a1ea-be6689e607ed","Type":"ContainerDied","Data":"0cf96799e07035171b9c610fea6313b824d405d222f18bc02ec5266501087f96"} Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.158672 4808 scope.go:117] "RemoveContainer" containerID="516b218bf929f18eb0453704db63a9a632bf10bea5cd9e6b24030634967a64fa" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.164221 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"99aebb2e-33d8-4657-9ed2-cf23aeb8a281","Type":"ContainerStarted","Data":"9ceabf0844087365c08ba52bc044c8e5bb247b8d65fb479bb3cd3d98e46721f8"} Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.192352 4808 scope.go:117] "RemoveContainer" containerID="40f7dc7f29d9aa9488c1ee47ca18e78ac062b2d82976a34fcdbda5c0ee3df075" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.211840 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.232462 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.251242 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 17:01:38 crc kubenswrapper[4808]: E1002 17:01:38.251719 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b409b3d8-0ed6-49e9-a1ea-be6689e607ed" containerName="rabbitmq" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.251743 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b409b3d8-0ed6-49e9-a1ea-be6689e607ed" containerName="rabbitmq" Oct 02 17:01:38 crc kubenswrapper[4808]: E1002 17:01:38.251773 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b409b3d8-0ed6-49e9-a1ea-be6689e607ed" containerName="setup-container" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.251784 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b409b3d8-0ed6-49e9-a1ea-be6689e607ed" containerName="setup-container" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.252021 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="b409b3d8-0ed6-49e9-a1ea-be6689e607ed" containerName="rabbitmq" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.253308 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.256815 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.257088 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.257238 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.257495 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-7qq44" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.257881 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.262225 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.262532 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.268890 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.349056 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/49331c3b-ac26-4f45-ad92-fadf45724c42-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.349146 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49331c3b-ac26-4f45-ad92-fadf45724c42-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.349180 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/49331c3b-ac26-4f45-ad92-fadf45724c42-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.349206 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.349290 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/49331c3b-ac26-4f45-ad92-fadf45724c42-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.349329 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/49331c3b-ac26-4f45-ad92-fadf45724c42-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.349413 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/49331c3b-ac26-4f45-ad92-fadf45724c42-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.349442 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/49331c3b-ac26-4f45-ad92-fadf45724c42-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.349468 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/49331c3b-ac26-4f45-ad92-fadf45724c42-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.349492 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh8pn\" (UniqueName: \"kubernetes.io/projected/49331c3b-ac26-4f45-ad92-fadf45724c42-kube-api-access-jh8pn\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.349613 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/49331c3b-ac26-4f45-ad92-fadf45724c42-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.451121 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/49331c3b-ac26-4f45-ad92-fadf45724c42-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.451238 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49331c3b-ac26-4f45-ad92-fadf45724c42-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.451295 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/49331c3b-ac26-4f45-ad92-fadf45724c42-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.451319 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.451342 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/49331c3b-ac26-4f45-ad92-fadf45724c42-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.451381 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/49331c3b-ac26-4f45-ad92-fadf45724c42-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.451460 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/49331c3b-ac26-4f45-ad92-fadf45724c42-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.451490 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/49331c3b-ac26-4f45-ad92-fadf45724c42-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.451526 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/49331c3b-ac26-4f45-ad92-fadf45724c42-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.451551 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh8pn\" (UniqueName: \"kubernetes.io/projected/49331c3b-ac26-4f45-ad92-fadf45724c42-kube-api-access-jh8pn\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.451586 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/49331c3b-ac26-4f45-ad92-fadf45724c42-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.451892 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/49331c3b-ac26-4f45-ad92-fadf45724c42-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.452130 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.452754 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/49331c3b-ac26-4f45-ad92-fadf45724c42-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.454033 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/49331c3b-ac26-4f45-ad92-fadf45724c42-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.454342 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/49331c3b-ac26-4f45-ad92-fadf45724c42-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.454362 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49331c3b-ac26-4f45-ad92-fadf45724c42-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.458402 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/49331c3b-ac26-4f45-ad92-fadf45724c42-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.466059 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/49331c3b-ac26-4f45-ad92-fadf45724c42-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.467445 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/49331c3b-ac26-4f45-ad92-fadf45724c42-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.474610 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/49331c3b-ac26-4f45-ad92-fadf45724c42-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.484046 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.492033 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh8pn\" (UniqueName: \"kubernetes.io/projected/49331c3b-ac26-4f45-ad92-fadf45724c42-kube-api-access-jh8pn\") pod \"rabbitmq-cell1-server-0\" (UID: \"49331c3b-ac26-4f45-ad92-fadf45724c42\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:38 crc kubenswrapper[4808]: I1002 17:01:38.634982 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:01:39 crc kubenswrapper[4808]: I1002 17:01:39.148720 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 17:01:39 crc kubenswrapper[4808]: W1002 17:01:39.156765 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49331c3b_ac26_4f45_ad92_fadf45724c42.slice/crio-ffcd40ad06ed9f3b779bab16827df34ebde28fa8df7eeea8d7074384500342d3 WatchSource:0}: Error finding container ffcd40ad06ed9f3b779bab16827df34ebde28fa8df7eeea8d7074384500342d3: Status 404 returned error can't find the container with id ffcd40ad06ed9f3b779bab16827df34ebde28fa8df7eeea8d7074384500342d3 Oct 02 17:01:39 crc kubenswrapper[4808]: I1002 17:01:39.175746 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"49331c3b-ac26-4f45-ad92-fadf45724c42","Type":"ContainerStarted","Data":"ffcd40ad06ed9f3b779bab16827df34ebde28fa8df7eeea8d7074384500342d3"} Oct 02 17:01:39 crc kubenswrapper[4808]: I1002 17:01:39.412532 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b409b3d8-0ed6-49e9-a1ea-be6689e607ed" path="/var/lib/kubelet/pods/b409b3d8-0ed6-49e9-a1ea-be6689e607ed/volumes" Oct 02 17:01:40 crc kubenswrapper[4808]: I1002 17:01:40.189593 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"99aebb2e-33d8-4657-9ed2-cf23aeb8a281","Type":"ContainerStarted","Data":"0b3592a95589c69639c6ab2212882946bc7b60fe2819dd879088e33a9084166f"} Oct 02 17:01:42 crc kubenswrapper[4808]: I1002 17:01:42.215566 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"49331c3b-ac26-4f45-ad92-fadf45724c42","Type":"ContainerStarted","Data":"cd9977ed50339c06922b539080e94153b91b8c2f34e1f133674f787cbcbcc67a"} Oct 02 17:01:42 crc kubenswrapper[4808]: I1002 17:01:42.959894 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-6wf8w"] Oct 02 17:01:42 crc kubenswrapper[4808]: I1002 17:01:42.964342 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:42 crc kubenswrapper[4808]: I1002 17:01:42.971324 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 02 17:01:42 crc kubenswrapper[4808]: I1002 17:01:42.977276 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-6wf8w"] Oct 02 17:01:43 crc kubenswrapper[4808]: I1002 17:01:43.144265 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-ovsdbserver-sb\") pod \"dnsmasq-dns-6447ccbd8f-6wf8w\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:43 crc kubenswrapper[4808]: I1002 17:01:43.144364 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-ovsdbserver-nb\") pod \"dnsmasq-dns-6447ccbd8f-6wf8w\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:43 crc kubenswrapper[4808]: I1002 17:01:43.144445 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-openstack-edpm-ipam\") pod \"dnsmasq-dns-6447ccbd8f-6wf8w\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:43 crc kubenswrapper[4808]: I1002 17:01:43.144478 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-dns-svc\") pod \"dnsmasq-dns-6447ccbd8f-6wf8w\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:43 crc kubenswrapper[4808]: I1002 17:01:43.144516 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-config\") pod \"dnsmasq-dns-6447ccbd8f-6wf8w\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:43 crc kubenswrapper[4808]: I1002 17:01:43.144566 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ft84\" (UniqueName: \"kubernetes.io/projected/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-kube-api-access-6ft84\") pod \"dnsmasq-dns-6447ccbd8f-6wf8w\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:43 crc kubenswrapper[4808]: I1002 17:01:43.245932 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-openstack-edpm-ipam\") pod \"dnsmasq-dns-6447ccbd8f-6wf8w\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:43 crc kubenswrapper[4808]: I1002 17:01:43.246032 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-dns-svc\") pod \"dnsmasq-dns-6447ccbd8f-6wf8w\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:43 crc kubenswrapper[4808]: I1002 17:01:43.246097 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-config\") pod \"dnsmasq-dns-6447ccbd8f-6wf8w\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:43 crc kubenswrapper[4808]: I1002 17:01:43.246181 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ft84\" (UniqueName: \"kubernetes.io/projected/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-kube-api-access-6ft84\") pod \"dnsmasq-dns-6447ccbd8f-6wf8w\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:43 crc kubenswrapper[4808]: I1002 17:01:43.246228 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-ovsdbserver-sb\") pod \"dnsmasq-dns-6447ccbd8f-6wf8w\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:43 crc kubenswrapper[4808]: I1002 17:01:43.246366 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-ovsdbserver-nb\") pod \"dnsmasq-dns-6447ccbd8f-6wf8w\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:43 crc kubenswrapper[4808]: I1002 17:01:43.248107 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-openstack-edpm-ipam\") pod \"dnsmasq-dns-6447ccbd8f-6wf8w\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:43 crc kubenswrapper[4808]: I1002 17:01:43.248516 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-ovsdbserver-sb\") pod \"dnsmasq-dns-6447ccbd8f-6wf8w\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:43 crc kubenswrapper[4808]: I1002 17:01:43.248668 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-dns-svc\") pod \"dnsmasq-dns-6447ccbd8f-6wf8w\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:43 crc kubenswrapper[4808]: I1002 17:01:43.249299 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-config\") pod \"dnsmasq-dns-6447ccbd8f-6wf8w\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:43 crc kubenswrapper[4808]: I1002 17:01:43.249374 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-ovsdbserver-nb\") pod \"dnsmasq-dns-6447ccbd8f-6wf8w\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:43 crc kubenswrapper[4808]: I1002 17:01:43.276984 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ft84\" (UniqueName: \"kubernetes.io/projected/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-kube-api-access-6ft84\") pod \"dnsmasq-dns-6447ccbd8f-6wf8w\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:43 crc kubenswrapper[4808]: I1002 17:01:43.295636 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:43 crc kubenswrapper[4808]: I1002 17:01:43.840193 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-6wf8w"] Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.135609 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897"] Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.138818 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.140503 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6w9jl" Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.140840 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.141626 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.147723 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.152362 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897"] Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.235391 4808 generic.go:334] "Generic (PLEG): container finished" podID="4ca877e1-cf28-465e-91d9-66a3b79bc3d5" containerID="73549811389c5cc5074d1f2c544a8446c5bbdf656ad91fa8dcdce669082a7bbb" exitCode=0 Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.235444 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" event={"ID":"4ca877e1-cf28-465e-91d9-66a3b79bc3d5","Type":"ContainerDied","Data":"73549811389c5cc5074d1f2c544a8446c5bbdf656ad91fa8dcdce669082a7bbb"} Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.235674 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" event={"ID":"4ca877e1-cf28-465e-91d9-66a3b79bc3d5","Type":"ContainerStarted","Data":"7e7ed2a249c97d465385d1200d7eafdb0348e46070ee9c25133551fcafe2f8b8"} Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.267349 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c04a4cc7-6d1c-419e-9751-94b14f873801-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gd897\" (UID: \"c04a4cc7-6d1c-419e-9751-94b14f873801\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.267465 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qmbg\" (UniqueName: \"kubernetes.io/projected/c04a4cc7-6d1c-419e-9751-94b14f873801-kube-api-access-9qmbg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gd897\" (UID: \"c04a4cc7-6d1c-419e-9751-94b14f873801\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.267567 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c04a4cc7-6d1c-419e-9751-94b14f873801-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gd897\" (UID: \"c04a4cc7-6d1c-419e-9751-94b14f873801\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.267595 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c04a4cc7-6d1c-419e-9751-94b14f873801-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gd897\" (UID: \"c04a4cc7-6d1c-419e-9751-94b14f873801\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.369052 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c04a4cc7-6d1c-419e-9751-94b14f873801-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gd897\" (UID: \"c04a4cc7-6d1c-419e-9751-94b14f873801\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.369653 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c04a4cc7-6d1c-419e-9751-94b14f873801-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gd897\" (UID: \"c04a4cc7-6d1c-419e-9751-94b14f873801\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.369817 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qmbg\" (UniqueName: \"kubernetes.io/projected/c04a4cc7-6d1c-419e-9751-94b14f873801-kube-api-access-9qmbg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gd897\" (UID: \"c04a4cc7-6d1c-419e-9751-94b14f873801\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.370344 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c04a4cc7-6d1c-419e-9751-94b14f873801-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gd897\" (UID: \"c04a4cc7-6d1c-419e-9751-94b14f873801\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.373645 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c04a4cc7-6d1c-419e-9751-94b14f873801-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gd897\" (UID: \"c04a4cc7-6d1c-419e-9751-94b14f873801\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.373766 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c04a4cc7-6d1c-419e-9751-94b14f873801-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gd897\" (UID: \"c04a4cc7-6d1c-419e-9751-94b14f873801\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.375064 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c04a4cc7-6d1c-419e-9751-94b14f873801-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gd897\" (UID: \"c04a4cc7-6d1c-419e-9751-94b14f873801\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.385182 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qmbg\" (UniqueName: \"kubernetes.io/projected/c04a4cc7-6d1c-419e-9751-94b14f873801-kube-api-access-9qmbg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gd897\" (UID: \"c04a4cc7-6d1c-419e-9751-94b14f873801\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" Oct 02 17:01:44 crc kubenswrapper[4808]: I1002 17:01:44.465067 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" Oct 02 17:01:45 crc kubenswrapper[4808]: W1002 17:01:45.013953 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc04a4cc7_6d1c_419e_9751_94b14f873801.slice/crio-43af45002cdab1703c5d3575f651491d58f89fdc3724364b9382f01c9eb5445c WatchSource:0}: Error finding container 43af45002cdab1703c5d3575f651491d58f89fdc3724364b9382f01c9eb5445c: Status 404 returned error can't find the container with id 43af45002cdab1703c5d3575f651491d58f89fdc3724364b9382f01c9eb5445c Oct 02 17:01:45 crc kubenswrapper[4808]: I1002 17:01:45.014947 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897"] Oct 02 17:01:45 crc kubenswrapper[4808]: I1002 17:01:45.248703 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" event={"ID":"4ca877e1-cf28-465e-91d9-66a3b79bc3d5","Type":"ContainerStarted","Data":"8dab5a97243bd256b9a904a60f8b0cda940790401aaf78c3d77b2a42f81fdc58"} Oct 02 17:01:45 crc kubenswrapper[4808]: I1002 17:01:45.249303 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:45 crc kubenswrapper[4808]: I1002 17:01:45.252987 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" event={"ID":"c04a4cc7-6d1c-419e-9751-94b14f873801","Type":"ContainerStarted","Data":"43af45002cdab1703c5d3575f651491d58f89fdc3724364b9382f01c9eb5445c"} Oct 02 17:01:45 crc kubenswrapper[4808]: I1002 17:01:45.274582 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" podStartSLOduration=3.27455544 podStartE2EDuration="3.27455544s" podCreationTimestamp="2025-10-02 17:01:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 17:01:45.266635047 +0000 UTC m=+1272.592164047" watchObservedRunningTime="2025-10-02 17:01:45.27455544 +0000 UTC m=+1272.600084460" Oct 02 17:01:48 crc kubenswrapper[4808]: I1002 17:01:48.740346 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 17:01:48 crc kubenswrapper[4808]: I1002 17:01:48.740798 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 17:01:48 crc kubenswrapper[4808]: I1002 17:01:48.740874 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 17:01:48 crc kubenswrapper[4808]: I1002 17:01:48.742154 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"325982695f20979226cf6b1b8bfe2ce4f159293f5eb31e61d627777c496fdb01"} pod="openshift-machine-config-operator/machine-config-daemon-7z66r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 17:01:48 crc kubenswrapper[4808]: I1002 17:01:48.742330 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" containerID="cri-o://325982695f20979226cf6b1b8bfe2ce4f159293f5eb31e61d627777c496fdb01" gracePeriod=600 Oct 02 17:01:53 crc kubenswrapper[4808]: I1002 17:01:53.297546 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:01:53 crc kubenswrapper[4808]: I1002 17:01:53.385567 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-7rnjh"] Oct 02 17:01:53 crc kubenswrapper[4808]: I1002 17:01:53.385923 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" podUID="e11ba3da-0594-456b-8cb1-2837d0422c2b" containerName="dnsmasq-dns" containerID="cri-o://61c2aadd62ddfbaaf03579537207f98e73b0f61d970ecc126ce0fb618e98079f" gracePeriod=10 Oct 02 17:01:55 crc kubenswrapper[4808]: I1002 17:01:55.905710 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5jdsw/must-gather-22xjr"] Oct 02 17:01:55 crc kubenswrapper[4808]: I1002 17:01:55.907496 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jdsw/must-gather-22xjr" Oct 02 17:01:55 crc kubenswrapper[4808]: I1002 17:01:55.910022 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-5jdsw"/"openshift-service-ca.crt" Oct 02 17:01:55 crc kubenswrapper[4808]: I1002 17:01:55.910724 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-5jdsw"/"kube-root-ca.crt" Oct 02 17:01:55 crc kubenswrapper[4808]: I1002 17:01:55.973178 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-5jdsw/must-gather-22xjr"] Oct 02 17:01:56 crc kubenswrapper[4808]: I1002 17:01:56.005314 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rs4w\" (UniqueName: \"kubernetes.io/projected/fb583e21-7f4e-40c6-b028-5209db1f4603-kube-api-access-2rs4w\") pod \"must-gather-22xjr\" (UID: \"fb583e21-7f4e-40c6-b028-5209db1f4603\") " pod="openshift-must-gather-5jdsw/must-gather-22xjr" Oct 02 17:01:56 crc kubenswrapper[4808]: I1002 17:01:56.005394 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fb583e21-7f4e-40c6-b028-5209db1f4603-must-gather-output\") pod \"must-gather-22xjr\" (UID: \"fb583e21-7f4e-40c6-b028-5209db1f4603\") " pod="openshift-must-gather-5jdsw/must-gather-22xjr" Oct 02 17:01:56 crc kubenswrapper[4808]: I1002 17:01:56.107526 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rs4w\" (UniqueName: \"kubernetes.io/projected/fb583e21-7f4e-40c6-b028-5209db1f4603-kube-api-access-2rs4w\") pod \"must-gather-22xjr\" (UID: \"fb583e21-7f4e-40c6-b028-5209db1f4603\") " pod="openshift-must-gather-5jdsw/must-gather-22xjr" Oct 02 17:01:56 crc kubenswrapper[4808]: I1002 17:01:56.107638 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fb583e21-7f4e-40c6-b028-5209db1f4603-must-gather-output\") pod \"must-gather-22xjr\" (UID: \"fb583e21-7f4e-40c6-b028-5209db1f4603\") " pod="openshift-must-gather-5jdsw/must-gather-22xjr" Oct 02 17:01:56 crc kubenswrapper[4808]: I1002 17:01:56.108288 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fb583e21-7f4e-40c6-b028-5209db1f4603-must-gather-output\") pod \"must-gather-22xjr\" (UID: \"fb583e21-7f4e-40c6-b028-5209db1f4603\") " pod="openshift-must-gather-5jdsw/must-gather-22xjr" Oct 02 17:01:56 crc kubenswrapper[4808]: I1002 17:01:56.129271 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rs4w\" (UniqueName: \"kubernetes.io/projected/fb583e21-7f4e-40c6-b028-5209db1f4603-kube-api-access-2rs4w\") pod \"must-gather-22xjr\" (UID: \"fb583e21-7f4e-40c6-b028-5209db1f4603\") " pod="openshift-must-gather-5jdsw/must-gather-22xjr" Oct 02 17:01:56 crc kubenswrapper[4808]: I1002 17:01:56.266196 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jdsw/must-gather-22xjr" Oct 02 17:01:56 crc kubenswrapper[4808]: I1002 17:01:56.715097 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-5jdsw/must-gather-22xjr"] Oct 02 17:01:57 crc kubenswrapper[4808]: I1002 17:01:57.104611 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-config-operator_machine-config-daemon-7z66r_f83ce425-101d-4489-94a4-5c256eb29328/machine-config-daemon/5.log" Oct 02 17:01:57 crc kubenswrapper[4808]: I1002 17:01:57.105510 4808 generic.go:334] "Generic (PLEG): container finished" podID="f83ce425-101d-4489-94a4-5c256eb29328" containerID="325982695f20979226cf6b1b8bfe2ce4f159293f5eb31e61d627777c496fdb01" exitCode=-1 Oct 02 17:01:57 crc kubenswrapper[4808]: I1002 17:01:57.105543 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" event={"ID":"f83ce425-101d-4489-94a4-5c256eb29328","Type":"ContainerDied","Data":"325982695f20979226cf6b1b8bfe2ce4f159293f5eb31e61d627777c496fdb01"} Oct 02 17:01:57 crc kubenswrapper[4808]: I1002 17:01:57.105599 4808 scope.go:117] "RemoveContainer" containerID="2d9c8f9e3115f67c6ef88d1a8e4688cf6f5cba6faca4d6c650065b45bf9b7b1b" Oct 02 17:01:57 crc kubenswrapper[4808]: I1002 17:01:57.327068 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" podUID="e11ba3da-0594-456b-8cb1-2837d0422c2b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.184:5353: connect: connection refused" Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.136090 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jdsw/must-gather-22xjr" event={"ID":"fb583e21-7f4e-40c6-b028-5209db1f4603","Type":"ContainerStarted","Data":"178860439a220dad0799f34359f581ca414a787d04eb481f9e5ac995db5155d3"} Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.576575 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-ptlg6"] Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.578600 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.658474 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-ptlg6"] Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.669989 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d411d4ec-d70f-4c68-9ebb-80fc8344e6bb-ovsdbserver-sb\") pod \"dnsmasq-dns-864d5fc68c-ptlg6\" (UID: \"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb\") " pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.670057 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d411d4ec-d70f-4c68-9ebb-80fc8344e6bb-dns-svc\") pod \"dnsmasq-dns-864d5fc68c-ptlg6\" (UID: \"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb\") " pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.670090 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7dbr\" (UniqueName: \"kubernetes.io/projected/d411d4ec-d70f-4c68-9ebb-80fc8344e6bb-kube-api-access-j7dbr\") pod \"dnsmasq-dns-864d5fc68c-ptlg6\" (UID: \"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb\") " pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.670130 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d411d4ec-d70f-4c68-9ebb-80fc8344e6bb-ovsdbserver-nb\") pod \"dnsmasq-dns-864d5fc68c-ptlg6\" (UID: \"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb\") " pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.670161 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d411d4ec-d70f-4c68-9ebb-80fc8344e6bb-openstack-edpm-ipam\") pod \"dnsmasq-dns-864d5fc68c-ptlg6\" (UID: \"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb\") " pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.670197 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d411d4ec-d70f-4c68-9ebb-80fc8344e6bb-config\") pod \"dnsmasq-dns-864d5fc68c-ptlg6\" (UID: \"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb\") " pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.771816 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d411d4ec-d70f-4c68-9ebb-80fc8344e6bb-ovsdbserver-sb\") pod \"dnsmasq-dns-864d5fc68c-ptlg6\" (UID: \"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb\") " pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.771901 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d411d4ec-d70f-4c68-9ebb-80fc8344e6bb-dns-svc\") pod \"dnsmasq-dns-864d5fc68c-ptlg6\" (UID: \"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb\") " pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.771926 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7dbr\" (UniqueName: \"kubernetes.io/projected/d411d4ec-d70f-4c68-9ebb-80fc8344e6bb-kube-api-access-j7dbr\") pod \"dnsmasq-dns-864d5fc68c-ptlg6\" (UID: \"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb\") " pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.771955 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d411d4ec-d70f-4c68-9ebb-80fc8344e6bb-ovsdbserver-nb\") pod \"dnsmasq-dns-864d5fc68c-ptlg6\" (UID: \"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb\") " pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.771982 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d411d4ec-d70f-4c68-9ebb-80fc8344e6bb-openstack-edpm-ipam\") pod \"dnsmasq-dns-864d5fc68c-ptlg6\" (UID: \"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb\") " pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.772011 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d411d4ec-d70f-4c68-9ebb-80fc8344e6bb-config\") pod \"dnsmasq-dns-864d5fc68c-ptlg6\" (UID: \"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb\") " pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.772907 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d411d4ec-d70f-4c68-9ebb-80fc8344e6bb-ovsdbserver-sb\") pod \"dnsmasq-dns-864d5fc68c-ptlg6\" (UID: \"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb\") " pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.772941 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d411d4ec-d70f-4c68-9ebb-80fc8344e6bb-ovsdbserver-nb\") pod \"dnsmasq-dns-864d5fc68c-ptlg6\" (UID: \"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb\") " pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.772990 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d411d4ec-d70f-4c68-9ebb-80fc8344e6bb-config\") pod \"dnsmasq-dns-864d5fc68c-ptlg6\" (UID: \"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb\") " pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.773180 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d411d4ec-d70f-4c68-9ebb-80fc8344e6bb-openstack-edpm-ipam\") pod \"dnsmasq-dns-864d5fc68c-ptlg6\" (UID: \"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb\") " pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.773215 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d411d4ec-d70f-4c68-9ebb-80fc8344e6bb-dns-svc\") pod \"dnsmasq-dns-864d5fc68c-ptlg6\" (UID: \"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb\") " pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.798520 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7dbr\" (UniqueName: \"kubernetes.io/projected/d411d4ec-d70f-4c68-9ebb-80fc8344e6bb-kube-api-access-j7dbr\") pod \"dnsmasq-dns-864d5fc68c-ptlg6\" (UID: \"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb\") " pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:01:58 crc kubenswrapper[4808]: I1002 17:01:58.926665 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:01:59 crc kubenswrapper[4808]: I1002 17:01:59.252393 4808 generic.go:334] "Generic (PLEG): container finished" podID="e11ba3da-0594-456b-8cb1-2837d0422c2b" containerID="61c2aadd62ddfbaaf03579537207f98e73b0f61d970ecc126ce0fb618e98079f" exitCode=0 Oct 02 17:01:59 crc kubenswrapper[4808]: I1002 17:01:59.252543 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" event={"ID":"e11ba3da-0594-456b-8cb1-2837d0422c2b","Type":"ContainerDied","Data":"61c2aadd62ddfbaaf03579537207f98e73b0f61d970ecc126ce0fb618e98079f"} Oct 02 17:01:59 crc kubenswrapper[4808]: I1002 17:01:59.430563 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-ptlg6"] Oct 02 17:02:00 crc kubenswrapper[4808]: I1002 17:02:00.263004 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" event={"ID":"f83ce425-101d-4489-94a4-5c256eb29328","Type":"ContainerStarted","Data":"0fcb0d9bbc76e2f98d002f08b3996525b0e799fbb853c38130d2b118c79a3da9"} Oct 02 17:02:00 crc kubenswrapper[4808]: I1002 17:02:00.266202 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" event={"ID":"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb","Type":"ContainerStarted","Data":"8c29bedeb13936deb11d338999f2e58d1c80921f4cd6956707ce4680fdf2fa7c"} Oct 02 17:02:00 crc kubenswrapper[4808]: I1002 17:02:00.266247 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" event={"ID":"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb","Type":"ContainerStarted","Data":"04297c9432d87a5d42b6d6db610bdd8a11c37930d28ab053a506d5b057c2ab6f"} Oct 02 17:02:01 crc kubenswrapper[4808]: I1002 17:02:01.276286 4808 generic.go:334] "Generic (PLEG): container finished" podID="d411d4ec-d70f-4c68-9ebb-80fc8344e6bb" containerID="8c29bedeb13936deb11d338999f2e58d1c80921f4cd6956707ce4680fdf2fa7c" exitCode=0 Oct 02 17:02:01 crc kubenswrapper[4808]: I1002 17:02:01.276494 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" event={"ID":"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb","Type":"ContainerDied","Data":"8c29bedeb13936deb11d338999f2e58d1c80921f4cd6956707ce4680fdf2fa7c"} Oct 02 17:02:07 crc kubenswrapper[4808]: I1002 17:02:07.326583 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" podUID="e11ba3da-0594-456b-8cb1-2837d0422c2b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.184:5353: i/o timeout" Oct 02 17:02:07 crc kubenswrapper[4808]: E1002 17:02:07.699192 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest" Oct 02 17:02:07 crc kubenswrapper[4808]: E1002 17:02:07.699567 4808 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 02 17:02:07 crc kubenswrapper[4808]: container &Container{Name:repo-setup-edpm-deployment-openstack-edpm-ipam,Image:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,Command:[],Args:[ansible-runner run /runner -p playbook.yaml -i repo-setup-edpm-deployment-openstack-edpm-ipam],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ANSIBLE_VERBOSITY,Value:2,ValueFrom:nil,},EnvVar{Name:RUNNER_PLAYBOOK,Value: Oct 02 17:02:07 crc kubenswrapper[4808]: - hosts: all Oct 02 17:02:07 crc kubenswrapper[4808]: strategy: linear Oct 02 17:02:07 crc kubenswrapper[4808]: tasks: Oct 02 17:02:07 crc kubenswrapper[4808]: - name: Enable podified-repos Oct 02 17:02:07 crc kubenswrapper[4808]: become: true Oct 02 17:02:07 crc kubenswrapper[4808]: ansible.builtin.shell: | Oct 02 17:02:07 crc kubenswrapper[4808]: set -euxo pipefail Oct 02 17:02:07 crc kubenswrapper[4808]: pushd /var/tmp Oct 02 17:02:07 crc kubenswrapper[4808]: curl -sL https://github.com/openstack-k8s-operators/repo-setup/archive/refs/heads/main.tar.gz | tar -xz Oct 02 17:02:07 crc kubenswrapper[4808]: pushd repo-setup-main Oct 02 17:02:07 crc kubenswrapper[4808]: python3 -m venv ./venv Oct 02 17:02:07 crc kubenswrapper[4808]: PBR_VERSION=0.0.0 ./venv/bin/pip install ./ Oct 02 17:02:07 crc kubenswrapper[4808]: ./venv/bin/repo-setup current-podified -b antelope Oct 02 17:02:07 crc kubenswrapper[4808]: popd Oct 02 17:02:07 crc kubenswrapper[4808]: rm -rf repo-setup-main Oct 02 17:02:07 crc kubenswrapper[4808]: Oct 02 17:02:07 crc kubenswrapper[4808]: Oct 02 17:02:07 crc kubenswrapper[4808]: ,ValueFrom:nil,},EnvVar{Name:RUNNER_EXTRA_VARS,Value: Oct 02 17:02:07 crc kubenswrapper[4808]: edpm_override_hosts: openstack-edpm-ipam Oct 02 17:02:07 crc kubenswrapper[4808]: edpm_service_type: repo-setup Oct 02 17:02:07 crc kubenswrapper[4808]: Oct 02 17:02:07 crc kubenswrapper[4808]: Oct 02 17:02:07 crc kubenswrapper[4808]: ,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:repo-setup-combined-ca-bundle,ReadOnly:false,MountPath:/var/lib/openstack/cacerts/repo-setup,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/runner/env/ssh_key,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:inventory,ReadOnly:false,MountPath:/runner/inventory/hosts,SubPath:inventory,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9qmbg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:openstack-aee-default-env,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod repo-setup-edpm-deployment-openstack-edpm-ipam-gd897_openstack(c04a4cc7-6d1c-419e-9751-94b14f873801): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Oct 02 17:02:07 crc kubenswrapper[4808]: > logger="UnhandledError" Oct 02 17:02:07 crc kubenswrapper[4808]: E1002 17:02:07.702315 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" podUID="c04a4cc7-6d1c-419e-9751-94b14f873801" Oct 02 17:02:07 crc kubenswrapper[4808]: I1002 17:02:07.853476 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:02:07 crc kubenswrapper[4808]: I1002 17:02:07.952095 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-config\") pod \"e11ba3da-0594-456b-8cb1-2837d0422c2b\" (UID: \"e11ba3da-0594-456b-8cb1-2837d0422c2b\") " Oct 02 17:02:07 crc kubenswrapper[4808]: I1002 17:02:07.952285 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-ovsdbserver-sb\") pod \"e11ba3da-0594-456b-8cb1-2837d0422c2b\" (UID: \"e11ba3da-0594-456b-8cb1-2837d0422c2b\") " Oct 02 17:02:07 crc kubenswrapper[4808]: I1002 17:02:07.952991 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6snvb\" (UniqueName: \"kubernetes.io/projected/e11ba3da-0594-456b-8cb1-2837d0422c2b-kube-api-access-6snvb\") pod \"e11ba3da-0594-456b-8cb1-2837d0422c2b\" (UID: \"e11ba3da-0594-456b-8cb1-2837d0422c2b\") " Oct 02 17:02:07 crc kubenswrapper[4808]: I1002 17:02:07.953132 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-ovsdbserver-nb\") pod \"e11ba3da-0594-456b-8cb1-2837d0422c2b\" (UID: \"e11ba3da-0594-456b-8cb1-2837d0422c2b\") " Oct 02 17:02:07 crc kubenswrapper[4808]: I1002 17:02:07.953301 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-dns-svc\") pod \"e11ba3da-0594-456b-8cb1-2837d0422c2b\" (UID: \"e11ba3da-0594-456b-8cb1-2837d0422c2b\") " Oct 02 17:02:07 crc kubenswrapper[4808]: I1002 17:02:07.958492 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e11ba3da-0594-456b-8cb1-2837d0422c2b-kube-api-access-6snvb" (OuterVolumeSpecName: "kube-api-access-6snvb") pod "e11ba3da-0594-456b-8cb1-2837d0422c2b" (UID: "e11ba3da-0594-456b-8cb1-2837d0422c2b"). InnerVolumeSpecName "kube-api-access-6snvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:02:08 crc kubenswrapper[4808]: I1002 17:02:08.009182 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e11ba3da-0594-456b-8cb1-2837d0422c2b" (UID: "e11ba3da-0594-456b-8cb1-2837d0422c2b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 17:02:08 crc kubenswrapper[4808]: I1002 17:02:08.014143 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e11ba3da-0594-456b-8cb1-2837d0422c2b" (UID: "e11ba3da-0594-456b-8cb1-2837d0422c2b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 17:02:08 crc kubenswrapper[4808]: I1002 17:02:08.019975 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-config" (OuterVolumeSpecName: "config") pod "e11ba3da-0594-456b-8cb1-2837d0422c2b" (UID: "e11ba3da-0594-456b-8cb1-2837d0422c2b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 17:02:08 crc kubenswrapper[4808]: I1002 17:02:08.026424 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e11ba3da-0594-456b-8cb1-2837d0422c2b" (UID: "e11ba3da-0594-456b-8cb1-2837d0422c2b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 17:02:08 crc kubenswrapper[4808]: I1002 17:02:08.055216 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 17:02:08 crc kubenswrapper[4808]: I1002 17:02:08.055261 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-config\") on node \"crc\" DevicePath \"\"" Oct 02 17:02:08 crc kubenswrapper[4808]: I1002 17:02:08.055272 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 17:02:08 crc kubenswrapper[4808]: I1002 17:02:08.055283 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6snvb\" (UniqueName: \"kubernetes.io/projected/e11ba3da-0594-456b-8cb1-2837d0422c2b-kube-api-access-6snvb\") on node \"crc\" DevicePath \"\"" Oct 02 17:02:08 crc kubenswrapper[4808]: I1002 17:02:08.055292 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e11ba3da-0594-456b-8cb1-2837d0422c2b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 17:02:08 crc kubenswrapper[4808]: I1002 17:02:08.353148 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" Oct 02 17:02:08 crc kubenswrapper[4808]: I1002 17:02:08.353166 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" event={"ID":"e11ba3da-0594-456b-8cb1-2837d0422c2b","Type":"ContainerDied","Data":"ec3bef622f9667e59968ba9631c6c7c47607292849c8671b6bc952d00bad8cbc"} Oct 02 17:02:08 crc kubenswrapper[4808]: I1002 17:02:08.353220 4808 scope.go:117] "RemoveContainer" containerID="61c2aadd62ddfbaaf03579537207f98e73b0f61d970ecc126ce0fb618e98079f" Oct 02 17:02:08 crc kubenswrapper[4808]: I1002 17:02:08.355803 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" event={"ID":"d411d4ec-d70f-4c68-9ebb-80fc8344e6bb","Type":"ContainerStarted","Data":"76e6419dc1ccb36c69e805cd5316891e5afb9ac1b4fc93415c4e3cfee40d6d2a"} Oct 02 17:02:08 crc kubenswrapper[4808]: I1002 17:02:08.355896 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:02:08 crc kubenswrapper[4808]: E1002 17:02:08.357137 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest\\\"\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" podUID="c04a4cc7-6d1c-419e-9751-94b14f873801" Oct 02 17:02:08 crc kubenswrapper[4808]: I1002 17:02:08.381812 4808 scope.go:117] "RemoveContainer" containerID="665dd85ef26f170f6e0fc1a9cc8d4962dda354e1245fdf60001f55028451190e" Oct 02 17:02:08 crc kubenswrapper[4808]: I1002 17:02:08.397876 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" podStartSLOduration=10.397857758 podStartE2EDuration="10.397857758s" podCreationTimestamp="2025-10-02 17:01:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 17:02:08.390623434 +0000 UTC m=+1295.716152474" watchObservedRunningTime="2025-10-02 17:02:08.397857758 +0000 UTC m=+1295.723386758" Oct 02 17:02:08 crc kubenswrapper[4808]: I1002 17:02:08.423829 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-7rnjh"] Oct 02 17:02:08 crc kubenswrapper[4808]: I1002 17:02:08.430781 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-7rnjh"] Oct 02 17:02:09 crc kubenswrapper[4808]: I1002 17:02:09.405447 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e11ba3da-0594-456b-8cb1-2837d0422c2b" path="/var/lib/kubelet/pods/e11ba3da-0594-456b-8cb1-2837d0422c2b/volumes" Oct 02 17:02:12 crc kubenswrapper[4808]: I1002 17:02:12.327790 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b856c5697-7rnjh" podUID="e11ba3da-0594-456b-8cb1-2837d0422c2b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.184:5353: i/o timeout" Oct 02 17:02:12 crc kubenswrapper[4808]: I1002 17:02:12.403299 4808 generic.go:334] "Generic (PLEG): container finished" podID="99aebb2e-33d8-4657-9ed2-cf23aeb8a281" containerID="0b3592a95589c69639c6ab2212882946bc7b60fe2819dd879088e33a9084166f" exitCode=0 Oct 02 17:02:12 crc kubenswrapper[4808]: I1002 17:02:12.403352 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"99aebb2e-33d8-4657-9ed2-cf23aeb8a281","Type":"ContainerDied","Data":"0b3592a95589c69639c6ab2212882946bc7b60fe2819dd879088e33a9084166f"} Oct 02 17:02:12 crc kubenswrapper[4808]: I1002 17:02:12.407791 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jdsw/must-gather-22xjr" event={"ID":"fb583e21-7f4e-40c6-b028-5209db1f4603","Type":"ContainerStarted","Data":"85efa474b6eaea6575403321be15c31d92778050b859a1ae7d700e35eec1ead4"} Oct 02 17:02:12 crc kubenswrapper[4808]: I1002 17:02:12.407844 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jdsw/must-gather-22xjr" event={"ID":"fb583e21-7f4e-40c6-b028-5209db1f4603","Type":"ContainerStarted","Data":"3d3ea3025dbb30e17129dc61f4937dfd4445737ebefec1770b8ca9fe23dafff4"} Oct 02 17:02:12 crc kubenswrapper[4808]: I1002 17:02:12.482480 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-5jdsw/must-gather-22xjr" podStartSLOduration=2.779878489 podStartE2EDuration="17.482454542s" podCreationTimestamp="2025-10-02 17:01:55 +0000 UTC" firstStartedPulling="2025-10-02 17:01:56.710972801 +0000 UTC m=+1284.036501801" lastFinishedPulling="2025-10-02 17:02:11.413548814 +0000 UTC m=+1298.739077854" observedRunningTime="2025-10-02 17:02:12.465072386 +0000 UTC m=+1299.790601426" watchObservedRunningTime="2025-10-02 17:02:12.482454542 +0000 UTC m=+1299.807983572" Oct 02 17:02:13 crc kubenswrapper[4808]: I1002 17:02:13.416991 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"99aebb2e-33d8-4657-9ed2-cf23aeb8a281","Type":"ContainerStarted","Data":"9fe4eae18abee83a4882536011a40b34592869018ad17a61b0d4be53fce10582"} Oct 02 17:02:13 crc kubenswrapper[4808]: I1002 17:02:13.417585 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 02 17:02:13 crc kubenswrapper[4808]: I1002 17:02:13.437834 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.437811038 podStartE2EDuration="37.437811038s" podCreationTimestamp="2025-10-02 17:01:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 17:02:13.436348279 +0000 UTC m=+1300.761877309" watchObservedRunningTime="2025-10-02 17:02:13.437811038 +0000 UTC m=+1300.763340038" Oct 02 17:02:13 crc kubenswrapper[4808]: I1002 17:02:13.928460 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-864d5fc68c-ptlg6" Oct 02 17:02:13 crc kubenswrapper[4808]: I1002 17:02:13.998785 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-6wf8w"] Oct 02 17:02:13 crc kubenswrapper[4808]: I1002 17:02:13.999364 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" podUID="4ca877e1-cf28-465e-91d9-66a3b79bc3d5" containerName="dnsmasq-dns" containerID="cri-o://8dab5a97243bd256b9a904a60f8b0cda940790401aaf78c3d77b2a42f81fdc58" gracePeriod=10 Oct 02 17:02:14 crc kubenswrapper[4808]: I1002 17:02:14.437890 4808 generic.go:334] "Generic (PLEG): container finished" podID="49331c3b-ac26-4f45-ad92-fadf45724c42" containerID="cd9977ed50339c06922b539080e94153b91b8c2f34e1f133674f787cbcbcc67a" exitCode=0 Oct 02 17:02:14 crc kubenswrapper[4808]: I1002 17:02:14.438128 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"49331c3b-ac26-4f45-ad92-fadf45724c42","Type":"ContainerDied","Data":"cd9977ed50339c06922b539080e94153b91b8c2f34e1f133674f787cbcbcc67a"} Oct 02 17:02:14 crc kubenswrapper[4808]: I1002 17:02:14.440390 4808 generic.go:334] "Generic (PLEG): container finished" podID="4ca877e1-cf28-465e-91d9-66a3b79bc3d5" containerID="8dab5a97243bd256b9a904a60f8b0cda940790401aaf78c3d77b2a42f81fdc58" exitCode=0 Oct 02 17:02:14 crc kubenswrapper[4808]: I1002 17:02:14.441669 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" event={"ID":"4ca877e1-cf28-465e-91d9-66a3b79bc3d5","Type":"ContainerDied","Data":"8dab5a97243bd256b9a904a60f8b0cda940790401aaf78c3d77b2a42f81fdc58"} Oct 02 17:02:15 crc kubenswrapper[4808]: I1002 17:02:15.842816 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5jdsw/crc-debug-8w4dv"] Oct 02 17:02:15 crc kubenswrapper[4808]: E1002 17:02:15.844707 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e11ba3da-0594-456b-8cb1-2837d0422c2b" containerName="dnsmasq-dns" Oct 02 17:02:15 crc kubenswrapper[4808]: I1002 17:02:15.844728 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e11ba3da-0594-456b-8cb1-2837d0422c2b" containerName="dnsmasq-dns" Oct 02 17:02:15 crc kubenswrapper[4808]: E1002 17:02:15.844748 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e11ba3da-0594-456b-8cb1-2837d0422c2b" containerName="init" Oct 02 17:02:15 crc kubenswrapper[4808]: I1002 17:02:15.844755 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e11ba3da-0594-456b-8cb1-2837d0422c2b" containerName="init" Oct 02 17:02:15 crc kubenswrapper[4808]: I1002 17:02:15.845129 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e11ba3da-0594-456b-8cb1-2837d0422c2b" containerName="dnsmasq-dns" Oct 02 17:02:15 crc kubenswrapper[4808]: I1002 17:02:15.845928 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jdsw/crc-debug-8w4dv" Oct 02 17:02:15 crc kubenswrapper[4808]: I1002 17:02:15.848681 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-5jdsw"/"default-dockercfg-nm2z4" Oct 02 17:02:15 crc kubenswrapper[4808]: I1002 17:02:15.922308 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3c51fb4d-7c38-449f-8d50-6bb8f34823b1-host\") pod \"crc-debug-8w4dv\" (UID: \"3c51fb4d-7c38-449f-8d50-6bb8f34823b1\") " pod="openshift-must-gather-5jdsw/crc-debug-8w4dv" Oct 02 17:02:15 crc kubenswrapper[4808]: I1002 17:02:15.922382 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8cqq\" (UniqueName: \"kubernetes.io/projected/3c51fb4d-7c38-449f-8d50-6bb8f34823b1-kube-api-access-w8cqq\") pod \"crc-debug-8w4dv\" (UID: \"3c51fb4d-7c38-449f-8d50-6bb8f34823b1\") " pod="openshift-must-gather-5jdsw/crc-debug-8w4dv" Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.046260 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3c51fb4d-7c38-449f-8d50-6bb8f34823b1-host\") pod \"crc-debug-8w4dv\" (UID: \"3c51fb4d-7c38-449f-8d50-6bb8f34823b1\") " pod="openshift-must-gather-5jdsw/crc-debug-8w4dv" Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.046532 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8cqq\" (UniqueName: \"kubernetes.io/projected/3c51fb4d-7c38-449f-8d50-6bb8f34823b1-kube-api-access-w8cqq\") pod \"crc-debug-8w4dv\" (UID: \"3c51fb4d-7c38-449f-8d50-6bb8f34823b1\") " pod="openshift-must-gather-5jdsw/crc-debug-8w4dv" Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.046551 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3c51fb4d-7c38-449f-8d50-6bb8f34823b1-host\") pod \"crc-debug-8w4dv\" (UID: \"3c51fb4d-7c38-449f-8d50-6bb8f34823b1\") " pod="openshift-must-gather-5jdsw/crc-debug-8w4dv" Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.065985 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8cqq\" (UniqueName: \"kubernetes.io/projected/3c51fb4d-7c38-449f-8d50-6bb8f34823b1-kube-api-access-w8cqq\") pod \"crc-debug-8w4dv\" (UID: \"3c51fb4d-7c38-449f-8d50-6bb8f34823b1\") " pod="openshift-must-gather-5jdsw/crc-debug-8w4dv" Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.164283 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jdsw/crc-debug-8w4dv" Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.460006 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"49331c3b-ac26-4f45-ad92-fadf45724c42","Type":"ContainerStarted","Data":"2160bb6fe6fb9ca01733552a408c4e753642b16f2887faf551a9fd931b0f71fc"} Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.466647 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jdsw/crc-debug-8w4dv" event={"ID":"3c51fb4d-7c38-449f-8d50-6bb8f34823b1","Type":"ContainerStarted","Data":"91cd411289a8626c96347b5f2e2c20c8fac8e0b2c647b0c0c4924f5056271f3b"} Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.774102 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.860457 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-config\") pod \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.860542 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-ovsdbserver-sb\") pod \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.860567 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-dns-svc\") pod \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.860622 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-ovsdbserver-nb\") pod \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.860706 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ft84\" (UniqueName: \"kubernetes.io/projected/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-kube-api-access-6ft84\") pod \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.860728 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-openstack-edpm-ipam\") pod \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\" (UID: \"4ca877e1-cf28-465e-91d9-66a3b79bc3d5\") " Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.884616 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-kube-api-access-6ft84" (OuterVolumeSpecName: "kube-api-access-6ft84") pod "4ca877e1-cf28-465e-91d9-66a3b79bc3d5" (UID: "4ca877e1-cf28-465e-91d9-66a3b79bc3d5"). InnerVolumeSpecName "kube-api-access-6ft84". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.928120 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4ca877e1-cf28-465e-91d9-66a3b79bc3d5" (UID: "4ca877e1-cf28-465e-91d9-66a3b79bc3d5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.939702 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4ca877e1-cf28-465e-91d9-66a3b79bc3d5" (UID: "4ca877e1-cf28-465e-91d9-66a3b79bc3d5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.940043 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-config" (OuterVolumeSpecName: "config") pod "4ca877e1-cf28-465e-91d9-66a3b79bc3d5" (UID: "4ca877e1-cf28-465e-91d9-66a3b79bc3d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.964473 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-config\") on node \"crc\" DevicePath \"\"" Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.964505 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.964517 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 17:02:16 crc kubenswrapper[4808]: I1002 17:02:16.964525 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ft84\" (UniqueName: \"kubernetes.io/projected/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-kube-api-access-6ft84\") on node \"crc\" DevicePath \"\"" Oct 02 17:02:17 crc kubenswrapper[4808]: I1002 17:02:17.008774 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4ca877e1-cf28-465e-91d9-66a3b79bc3d5" (UID: "4ca877e1-cf28-465e-91d9-66a3b79bc3d5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 17:02:17 crc kubenswrapper[4808]: I1002 17:02:17.025193 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "4ca877e1-cf28-465e-91d9-66a3b79bc3d5" (UID: "4ca877e1-cf28-465e-91d9-66a3b79bc3d5"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 17:02:17 crc kubenswrapper[4808]: I1002 17:02:17.066052 4808 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 02 17:02:17 crc kubenswrapper[4808]: I1002 17:02:17.066079 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ca877e1-cf28-465e-91d9-66a3b79bc3d5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 17:02:17 crc kubenswrapper[4808]: I1002 17:02:17.485784 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" event={"ID":"4ca877e1-cf28-465e-91d9-66a3b79bc3d5","Type":"ContainerDied","Data":"7e7ed2a249c97d465385d1200d7eafdb0348e46070ee9c25133551fcafe2f8b8"} Oct 02 17:02:17 crc kubenswrapper[4808]: I1002 17:02:17.485837 4808 scope.go:117] "RemoveContainer" containerID="8dab5a97243bd256b9a904a60f8b0cda940790401aaf78c3d77b2a42f81fdc58" Oct 02 17:02:17 crc kubenswrapper[4808]: I1002 17:02:17.485859 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-6wf8w" Oct 02 17:02:17 crc kubenswrapper[4808]: I1002 17:02:17.486288 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:02:17 crc kubenswrapper[4808]: I1002 17:02:17.510364 4808 scope.go:117] "RemoveContainer" containerID="73549811389c5cc5074d1f2c544a8446c5bbdf656ad91fa8dcdce669082a7bbb" Oct 02 17:02:17 crc kubenswrapper[4808]: I1002 17:02:17.516463 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=39.516446362 podStartE2EDuration="39.516446362s" podCreationTimestamp="2025-10-02 17:01:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 17:02:17.510007599 +0000 UTC m=+1304.835536599" watchObservedRunningTime="2025-10-02 17:02:17.516446362 +0000 UTC m=+1304.841975362" Oct 02 17:02:17 crc kubenswrapper[4808]: I1002 17:02:17.538730 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-6wf8w"] Oct 02 17:02:17 crc kubenswrapper[4808]: I1002 17:02:17.545422 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-6wf8w"] Oct 02 17:02:19 crc kubenswrapper[4808]: I1002 17:02:19.405524 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ca877e1-cf28-465e-91d9-66a3b79bc3d5" path="/var/lib/kubelet/pods/4ca877e1-cf28-465e-91d9-66a3b79bc3d5/volumes" Oct 02 17:02:26 crc kubenswrapper[4808]: I1002 17:02:26.845439 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 02 17:02:27 crc kubenswrapper[4808]: I1002 17:02:27.585363 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jdsw/crc-debug-8w4dv" event={"ID":"3c51fb4d-7c38-449f-8d50-6bb8f34823b1","Type":"ContainerStarted","Data":"2986dd928bde6cf69270dddd5c0f22363ef81625c628a0bcbea2b8e1878e72bc"} Oct 02 17:02:27 crc kubenswrapper[4808]: I1002 17:02:27.588821 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" event={"ID":"c04a4cc7-6d1c-419e-9751-94b14f873801","Type":"ContainerStarted","Data":"b1907517c02333b5cf10e2233f51f6cb68b654a9c51f8ca74dec62bec4139335"} Oct 02 17:02:27 crc kubenswrapper[4808]: I1002 17:02:27.599200 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-5jdsw/crc-debug-8w4dv" podStartSLOduration=2.3550273 podStartE2EDuration="12.599178259s" podCreationTimestamp="2025-10-02 17:02:15 +0000 UTC" firstStartedPulling="2025-10-02 17:02:16.206591482 +0000 UTC m=+1303.532120482" lastFinishedPulling="2025-10-02 17:02:26.450742431 +0000 UTC m=+1313.776271441" observedRunningTime="2025-10-02 17:02:27.596780485 +0000 UTC m=+1314.922309485" watchObservedRunningTime="2025-10-02 17:02:27.599178259 +0000 UTC m=+1314.924707269" Oct 02 17:02:27 crc kubenswrapper[4808]: I1002 17:02:27.618263 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" podStartSLOduration=2.22031132 podStartE2EDuration="43.618219499s" podCreationTimestamp="2025-10-02 17:01:44 +0000 UTC" firstStartedPulling="2025-10-02 17:01:45.016105679 +0000 UTC m=+1272.341634719" lastFinishedPulling="2025-10-02 17:02:26.414013898 +0000 UTC m=+1313.739542898" observedRunningTime="2025-10-02 17:02:27.610287627 +0000 UTC m=+1314.935816657" watchObservedRunningTime="2025-10-02 17:02:27.618219499 +0000 UTC m=+1314.943748499" Oct 02 17:02:28 crc kubenswrapper[4808]: I1002 17:02:28.637367 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 02 17:02:33 crc kubenswrapper[4808]: I1002 17:02:33.637277 4808 generic.go:334] "Generic (PLEG): container finished" podID="818e02a3-653a-4b36-a68b-9d4e2df0b71e" containerID="dc9639f10a3923cdab18975fdbea062b709d111e09fdf3caa99a2b5ab5ce0a27" exitCode=1 Oct 02 17:02:33 crc kubenswrapper[4808]: I1002 17:02:33.637445 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" event={"ID":"818e02a3-653a-4b36-a68b-9d4e2df0b71e","Type":"ContainerDied","Data":"dc9639f10a3923cdab18975fdbea062b709d111e09fdf3caa99a2b5ab5ce0a27"} Oct 02 17:02:33 crc kubenswrapper[4808]: I1002 17:02:33.637654 4808 scope.go:117] "RemoveContainer" containerID="a918ee069243e5dc0b7673d53697e23876162204008e5ae0c51065fa9012a203" Oct 02 17:02:33 crc kubenswrapper[4808]: I1002 17:02:33.638223 4808 scope.go:117] "RemoveContainer" containerID="dc9639f10a3923cdab18975fdbea062b709d111e09fdf3caa99a2b5ab5ce0a27" Oct 02 17:02:33 crc kubenswrapper[4808]: E1002 17:02:33.638440 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-65f4bcbb94-svkdd_openstack-operators(818e02a3-653a-4b36-a68b-9d4e2df0b71e)\"" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" podUID="818e02a3-653a-4b36-a68b-9d4e2df0b71e" Oct 02 17:02:34 crc kubenswrapper[4808]: I1002 17:02:34.725601 4808 scope.go:117] "RemoveContainer" containerID="b46fa00211cf2548288c47a6e3f0f2d480b75fcb70fd1028ae56f9d300ce5ecc" Oct 02 17:02:40 crc kubenswrapper[4808]: I1002 17:02:40.703449 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 17:02:40 crc kubenswrapper[4808]: I1002 17:02:40.704054 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 17:02:40 crc kubenswrapper[4808]: I1002 17:02:40.704863 4808 scope.go:117] "RemoveContainer" containerID="dc9639f10a3923cdab18975fdbea062b709d111e09fdf3caa99a2b5ab5ce0a27" Oct 02 17:02:40 crc kubenswrapper[4808]: E1002 17:02:40.705207 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-65f4bcbb94-svkdd_openstack-operators(818e02a3-653a-4b36-a68b-9d4e2df0b71e)\"" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" podUID="818e02a3-653a-4b36-a68b-9d4e2df0b71e" Oct 02 17:02:40 crc kubenswrapper[4808]: I1002 17:02:40.719653 4808 generic.go:334] "Generic (PLEG): container finished" podID="c04a4cc7-6d1c-419e-9751-94b14f873801" containerID="b1907517c02333b5cf10e2233f51f6cb68b654a9c51f8ca74dec62bec4139335" exitCode=0 Oct 02 17:02:40 crc kubenswrapper[4808]: I1002 17:02:40.719705 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" event={"ID":"c04a4cc7-6d1c-419e-9751-94b14f873801","Type":"ContainerDied","Data":"b1907517c02333b5cf10e2233f51f6cb68b654a9c51f8ca74dec62bec4139335"} Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.178703 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.359583 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c04a4cc7-6d1c-419e-9751-94b14f873801-ssh-key\") pod \"c04a4cc7-6d1c-419e-9751-94b14f873801\" (UID: \"c04a4cc7-6d1c-419e-9751-94b14f873801\") " Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.359703 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c04a4cc7-6d1c-419e-9751-94b14f873801-inventory\") pod \"c04a4cc7-6d1c-419e-9751-94b14f873801\" (UID: \"c04a4cc7-6d1c-419e-9751-94b14f873801\") " Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.359854 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qmbg\" (UniqueName: \"kubernetes.io/projected/c04a4cc7-6d1c-419e-9751-94b14f873801-kube-api-access-9qmbg\") pod \"c04a4cc7-6d1c-419e-9751-94b14f873801\" (UID: \"c04a4cc7-6d1c-419e-9751-94b14f873801\") " Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.359911 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c04a4cc7-6d1c-419e-9751-94b14f873801-repo-setup-combined-ca-bundle\") pod \"c04a4cc7-6d1c-419e-9751-94b14f873801\" (UID: \"c04a4cc7-6d1c-419e-9751-94b14f873801\") " Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.369791 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c04a4cc7-6d1c-419e-9751-94b14f873801-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "c04a4cc7-6d1c-419e-9751-94b14f873801" (UID: "c04a4cc7-6d1c-419e-9751-94b14f873801"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.371095 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c04a4cc7-6d1c-419e-9751-94b14f873801-kube-api-access-9qmbg" (OuterVolumeSpecName: "kube-api-access-9qmbg") pod "c04a4cc7-6d1c-419e-9751-94b14f873801" (UID: "c04a4cc7-6d1c-419e-9751-94b14f873801"). InnerVolumeSpecName "kube-api-access-9qmbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.423979 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c04a4cc7-6d1c-419e-9751-94b14f873801-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c04a4cc7-6d1c-419e-9751-94b14f873801" (UID: "c04a4cc7-6d1c-419e-9751-94b14f873801"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.446412 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c04a4cc7-6d1c-419e-9751-94b14f873801-inventory" (OuterVolumeSpecName: "inventory") pod "c04a4cc7-6d1c-419e-9751-94b14f873801" (UID: "c04a4cc7-6d1c-419e-9751-94b14f873801"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.462340 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c04a4cc7-6d1c-419e-9751-94b14f873801-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.462374 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c04a4cc7-6d1c-419e-9751-94b14f873801-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.462384 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qmbg\" (UniqueName: \"kubernetes.io/projected/c04a4cc7-6d1c-419e-9751-94b14f873801-kube-api-access-9qmbg\") on node \"crc\" DevicePath \"\"" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.462396 4808 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c04a4cc7-6d1c-419e-9751-94b14f873801-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.742870 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" event={"ID":"c04a4cc7-6d1c-419e-9751-94b14f873801","Type":"ContainerDied","Data":"43af45002cdab1703c5d3575f651491d58f89fdc3724364b9382f01c9eb5445c"} Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.742917 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43af45002cdab1703c5d3575f651491d58f89fdc3724364b9382f01c9eb5445c" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.742941 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gd897" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.821708 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9"] Oct 02 17:02:42 crc kubenswrapper[4808]: E1002 17:02:42.822369 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ca877e1-cf28-465e-91d9-66a3b79bc3d5" containerName="dnsmasq-dns" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.822398 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ca877e1-cf28-465e-91d9-66a3b79bc3d5" containerName="dnsmasq-dns" Oct 02 17:02:42 crc kubenswrapper[4808]: E1002 17:02:42.822470 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c04a4cc7-6d1c-419e-9751-94b14f873801" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.822494 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c04a4cc7-6d1c-419e-9751-94b14f873801" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 17:02:42 crc kubenswrapper[4808]: E1002 17:02:42.822536 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ca877e1-cf28-465e-91d9-66a3b79bc3d5" containerName="init" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.822550 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ca877e1-cf28-465e-91d9-66a3b79bc3d5" containerName="init" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.822883 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c04a4cc7-6d1c-419e-9751-94b14f873801" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.822934 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ca877e1-cf28-465e-91d9-66a3b79bc3d5" containerName="dnsmasq-dns" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.824017 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.826691 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.826802 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6w9jl" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.827321 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.829083 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.832958 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9"] Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.969303 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/40e44450-b30c-4df1-b375-ac76b8ae7719-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9\" (UID: \"40e44450-b30c-4df1-b375-ac76b8ae7719\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.969381 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/40e44450-b30c-4df1-b375-ac76b8ae7719-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9\" (UID: \"40e44450-b30c-4df1-b375-ac76b8ae7719\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.969501 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e44450-b30c-4df1-b375-ac76b8ae7719-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9\" (UID: \"40e44450-b30c-4df1-b375-ac76b8ae7719\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" Oct 02 17:02:42 crc kubenswrapper[4808]: I1002 17:02:42.969547 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txcsd\" (UniqueName: \"kubernetes.io/projected/40e44450-b30c-4df1-b375-ac76b8ae7719-kube-api-access-txcsd\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9\" (UID: \"40e44450-b30c-4df1-b375-ac76b8ae7719\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" Oct 02 17:02:43 crc kubenswrapper[4808]: I1002 17:02:43.070827 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e44450-b30c-4df1-b375-ac76b8ae7719-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9\" (UID: \"40e44450-b30c-4df1-b375-ac76b8ae7719\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" Oct 02 17:02:43 crc kubenswrapper[4808]: I1002 17:02:43.070896 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txcsd\" (UniqueName: \"kubernetes.io/projected/40e44450-b30c-4df1-b375-ac76b8ae7719-kube-api-access-txcsd\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9\" (UID: \"40e44450-b30c-4df1-b375-ac76b8ae7719\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" Oct 02 17:02:43 crc kubenswrapper[4808]: I1002 17:02:43.070935 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/40e44450-b30c-4df1-b375-ac76b8ae7719-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9\" (UID: \"40e44450-b30c-4df1-b375-ac76b8ae7719\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" Oct 02 17:02:43 crc kubenswrapper[4808]: I1002 17:02:43.070970 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/40e44450-b30c-4df1-b375-ac76b8ae7719-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9\" (UID: \"40e44450-b30c-4df1-b375-ac76b8ae7719\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" Oct 02 17:02:43 crc kubenswrapper[4808]: I1002 17:02:43.075954 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e44450-b30c-4df1-b375-ac76b8ae7719-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9\" (UID: \"40e44450-b30c-4df1-b375-ac76b8ae7719\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" Oct 02 17:02:43 crc kubenswrapper[4808]: I1002 17:02:43.076116 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/40e44450-b30c-4df1-b375-ac76b8ae7719-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9\" (UID: \"40e44450-b30c-4df1-b375-ac76b8ae7719\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" Oct 02 17:02:43 crc kubenswrapper[4808]: I1002 17:02:43.084763 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/40e44450-b30c-4df1-b375-ac76b8ae7719-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9\" (UID: \"40e44450-b30c-4df1-b375-ac76b8ae7719\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" Oct 02 17:02:43 crc kubenswrapper[4808]: I1002 17:02:43.092771 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txcsd\" (UniqueName: \"kubernetes.io/projected/40e44450-b30c-4df1-b375-ac76b8ae7719-kube-api-access-txcsd\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9\" (UID: \"40e44450-b30c-4df1-b375-ac76b8ae7719\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" Oct 02 17:02:43 crc kubenswrapper[4808]: I1002 17:02:43.142748 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" Oct 02 17:02:43 crc kubenswrapper[4808]: I1002 17:02:43.662539 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9"] Oct 02 17:02:43 crc kubenswrapper[4808]: I1002 17:02:43.751329 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" event={"ID":"40e44450-b30c-4df1-b375-ac76b8ae7719","Type":"ContainerStarted","Data":"b0ce94799e09ec7b1ecce9f6708ecc92a791fc3551641b6309c8966e91a9e547"} Oct 02 17:02:44 crc kubenswrapper[4808]: I1002 17:02:44.760554 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" event={"ID":"40e44450-b30c-4df1-b375-ac76b8ae7719","Type":"ContainerStarted","Data":"5bb64046286c402dd492725c73918916c21a02daf5f25fb1268b0a9a82f88d03"} Oct 02 17:02:44 crc kubenswrapper[4808]: I1002 17:02:44.775846 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" podStartSLOduration=2.316632823 podStartE2EDuration="2.775830022s" podCreationTimestamp="2025-10-02 17:02:42 +0000 UTC" firstStartedPulling="2025-10-02 17:02:43.679770607 +0000 UTC m=+1331.005299607" lastFinishedPulling="2025-10-02 17:02:44.138967806 +0000 UTC m=+1331.464496806" observedRunningTime="2025-10-02 17:02:44.774546198 +0000 UTC m=+1332.100075208" watchObservedRunningTime="2025-10-02 17:02:44.775830022 +0000 UTC m=+1332.101359022" Oct 02 17:02:52 crc kubenswrapper[4808]: I1002 17:02:52.396883 4808 scope.go:117] "RemoveContainer" containerID="dc9639f10a3923cdab18975fdbea062b709d111e09fdf3caa99a2b5ab5ce0a27" Oct 02 17:02:52 crc kubenswrapper[4808]: E1002 17:02:52.397978 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-65f4bcbb94-svkdd_openstack-operators(818e02a3-653a-4b36-a68b-9d4e2df0b71e)\"" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" podUID="818e02a3-653a-4b36-a68b-9d4e2df0b71e" Oct 02 17:03:04 crc kubenswrapper[4808]: I1002 17:03:04.403024 4808 scope.go:117] "RemoveContainer" containerID="dc9639f10a3923cdab18975fdbea062b709d111e09fdf3caa99a2b5ab5ce0a27" Oct 02 17:03:04 crc kubenswrapper[4808]: I1002 17:03:04.938150 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" event={"ID":"818e02a3-653a-4b36-a68b-9d4e2df0b71e","Type":"ContainerStarted","Data":"f38c1e1eaa31905554cb4c587be088935750e18481e081f3175807eb6c29368c"} Oct 02 17:03:04 crc kubenswrapper[4808]: I1002 17:03:04.938874 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 17:03:10 crc kubenswrapper[4808]: I1002 17:03:10.707664 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 17:03:14 crc kubenswrapper[4808]: I1002 17:03:14.860051 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-4027-account-create-77w26_7951034c-d159-4816-8f1e-4118953ac9f0/mariadb-account-create/0.log" Oct 02 17:03:15 crc kubenswrapper[4808]: I1002 17:03:15.066438 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5684d76658-g9f5f_f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9/barbican-api/0.log" Oct 02 17:03:15 crc kubenswrapper[4808]: I1002 17:03:15.228608 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5684d76658-g9f5f_f8a6f25d-f9a9-4fb4-8b97-c06144ea39b9/barbican-api-log/0.log" Oct 02 17:03:15 crc kubenswrapper[4808]: I1002 17:03:15.391785 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-db-create-t6tjw_f2120107-7ced-4921-9a2d-eba436f36928/mariadb-database-create/0.log" Oct 02 17:03:15 crc kubenswrapper[4808]: I1002 17:03:15.551806 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-db-sync-ckf4j_74a1c6ad-0412-4c46-90ce-34869c9ca9e6/barbican-db-sync/0.log" Oct 02 17:03:15 crc kubenswrapper[4808]: I1002 17:03:15.713865 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-54854ffb46-kktlh_f36f9202-6bbb-404f-865d-206a87585936/barbican-keystone-listener/0.log" Oct 02 17:03:15 crc kubenswrapper[4808]: I1002 17:03:15.806462 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-54854ffb46-kktlh_f36f9202-6bbb-404f-865d-206a87585936/barbican-keystone-listener-log/0.log" Oct 02 17:03:15 crc kubenswrapper[4808]: I1002 17:03:15.952461 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5b4f6b689c-zbtm2_695150b9-231f-4f3b-8e53-47b067a55025/barbican-worker/0.log" Oct 02 17:03:16 crc kubenswrapper[4808]: I1002 17:03:16.067626 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5b4f6b689c-zbtm2_695150b9-231f-4f3b-8e53-47b067a55025/barbican-worker-log/0.log" Oct 02 17:03:16 crc kubenswrapper[4808]: I1002 17:03:16.249009 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9_40e44450-b30c-4df1-b375-ac76b8ae7719/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 17:03:16 crc kubenswrapper[4808]: I1002 17:03:16.449032 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_33df7922-0de2-4897-9cfa-599fadc6229f/ceilometer-central-agent/0.log" Oct 02 17:03:16 crc kubenswrapper[4808]: I1002 17:03:16.551522 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_33df7922-0de2-4897-9cfa-599fadc6229f/ceilometer-notification-agent/0.log" Oct 02 17:03:16 crc kubenswrapper[4808]: I1002 17:03:16.633099 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_33df7922-0de2-4897-9cfa-599fadc6229f/proxy-httpd/0.log" Oct 02 17:03:16 crc kubenswrapper[4808]: I1002 17:03:16.744542 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_33df7922-0de2-4897-9cfa-599fadc6229f/sg-core/0.log" Oct 02 17:03:16 crc kubenswrapper[4808]: I1002 17:03:16.934204 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-7b56-account-create-5z4wr_f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a/mariadb-account-create/0.log" Oct 02 17:03:17 crc kubenswrapper[4808]: I1002 17:03:17.104345 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_7723b385-3e7c-4c10-bf48-f2213ce31f16/cinder-api/0.log" Oct 02 17:03:17 crc kubenswrapper[4808]: I1002 17:03:17.270049 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_7723b385-3e7c-4c10-bf48-f2213ce31f16/cinder-api-log/0.log" Oct 02 17:03:17 crc kubenswrapper[4808]: I1002 17:03:17.444664 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-db-create-rr4z9_af84bfb8-1583-46a8-ae8f-1b28b3568311/mariadb-database-create/0.log" Oct 02 17:03:17 crc kubenswrapper[4808]: I1002 17:03:17.541531 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-db-sync-l2fls_b3d4166a-c265-4cb4-8481-32e6b0144997/cinder-db-sync/0.log" Oct 02 17:03:17 crc kubenswrapper[4808]: I1002 17:03:17.705746 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_a608d71c-370e-46ca-bf97-1752628865ad/cinder-scheduler/0.log" Oct 02 17:03:17 crc kubenswrapper[4808]: I1002 17:03:17.723651 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_a608d71c-370e-46ca-bf97-1752628865ad/probe/0.log" Oct 02 17:03:17 crc kubenswrapper[4808]: I1002 17:03:17.885001 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-864d5fc68c-ptlg6_d411d4ec-d70f-4c68-9ebb-80fc8344e6bb/init/0.log" Oct 02 17:03:18 crc kubenswrapper[4808]: I1002 17:03:18.011607 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-864d5fc68c-ptlg6_d411d4ec-d70f-4c68-9ebb-80fc8344e6bb/init/0.log" Oct 02 17:03:18 crc kubenswrapper[4808]: I1002 17:03:18.065414 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-864d5fc68c-ptlg6_d411d4ec-d70f-4c68-9ebb-80fc8344e6bb/dnsmasq-dns/0.log" Oct 02 17:03:18 crc kubenswrapper[4808]: I1002 17:03:18.199051 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-b4e3-account-create-qsp8f_387ae2f5-c39e-44a8-b471-64ab91a40d01/mariadb-account-create/0.log" Oct 02 17:03:18 crc kubenswrapper[4808]: I1002 17:03:18.449036 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-db-create-rk9zb_16cb8757-fcb8-469a-8971-1c6dffa2e3fd/mariadb-database-create/0.log" Oct 02 17:03:18 crc kubenswrapper[4808]: I1002 17:03:18.596079 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-db-sync-67d46_8a5d0b28-6ba9-4421-98b6-1c6280ed7937/glance-db-sync/0.log" Oct 02 17:03:18 crc kubenswrapper[4808]: I1002 17:03:18.709192 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-3d12-account-create-wqs6d_f446b584-b5e2-48fa-9f7d-3b6ccdc92944/mariadb-account-create/0.log" Oct 02 17:03:18 crc kubenswrapper[4808]: I1002 17:03:18.894921 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-577d5bdf64-z4hwp_9300a6c4-e3c6-4cef-98bf-192b9d5549be/keystone-api/0.log" Oct 02 17:03:18 crc kubenswrapper[4808]: I1002 17:03:18.971848 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-bootstrap-ls5jx_60c1cab9-fa93-4572-8eb5-f9682a047645/keystone-bootstrap/0.log" Oct 02 17:03:19 crc kubenswrapper[4808]: I1002 17:03:19.109912 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29323741-x5mvf_765ddaf0-84a9-4ec2-b07e-683ebe340353/keystone-cron/0.log" Oct 02 17:03:19 crc kubenswrapper[4808]: I1002 17:03:19.252198 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-db-create-twjnb_06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c/mariadb-database-create/0.log" Oct 02 17:03:19 crc kubenswrapper[4808]: I1002 17:03:19.332960 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-db-sync-7wz2m_cfeb1719-f48e-488b-8c22-a9be946819e3/keystone-db-sync/0.log" Oct 02 17:03:19 crc kubenswrapper[4808]: I1002 17:03:19.457854 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_93c29105-8d7b-4104-9d99-b5e4bd697253/kube-state-metrics/0.log" Oct 02 17:03:19 crc kubenswrapper[4808]: I1002 17:03:19.718782 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-78b67777b5-h245x_72c31691-57ad-4fa3-9fad-b81d7655d443/neutron-api/0.log" Oct 02 17:03:19 crc kubenswrapper[4808]: I1002 17:03:19.869172 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-78b67777b5-h245x_72c31691-57ad-4fa3-9fad-b81d7655d443/neutron-httpd/0.log" Oct 02 17:03:20 crc kubenswrapper[4808]: I1002 17:03:20.069950 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7c90-account-create-z4xqz_afc592f9-999b-4787-ae91-bdcd10aa2ad5/mariadb-account-create/0.log" Oct 02 17:03:20 crc kubenswrapper[4808]: I1002 17:03:20.273258 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-db-create-6bh5l_9f95dff9-cca1-4356-912a-76d1475a273a/mariadb-database-create/0.log" Oct 02 17:03:20 crc kubenswrapper[4808]: I1002 17:03:20.545596 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-db-sync-zpkmw_c50129b6-292e-434b-a617-2781dac4101e/neutron-db-sync/0.log" Oct 02 17:03:20 crc kubenswrapper[4808]: I1002 17:03:20.795605 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_fa5606b5-d0de-4403-b815-41d3216754d5/nova-api-api/0.log" Oct 02 17:03:20 crc kubenswrapper[4808]: I1002 17:03:20.944545 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_fa5606b5-d0de-4403-b815-41d3216754d5/nova-api-log/0.log" Oct 02 17:03:21 crc kubenswrapper[4808]: I1002 17:03:21.101394 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-5e0c-account-create-75q46_5a2cb272-0bff-49dc-bf02-114dcfda4aea/mariadb-account-create/0.log" Oct 02 17:03:21 crc kubenswrapper[4808]: I1002 17:03:21.302627 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-db-create-2ws9d_91cad4f7-82d4-4b24-b77d-4e98526c6d77/mariadb-database-create/0.log" Oct 02 17:03:21 crc kubenswrapper[4808]: I1002 17:03:21.491637 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-731b-account-create-zrkfk_e94ceeb9-6ff7-4fcd-8727-64bb034bef2f/mariadb-account-create/0.log" Oct 02 17:03:21 crc kubenswrapper[4808]: I1002 17:03:21.663639 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-cell-mapping-x8vmm_c4bb01f7-7c71-4b46-b45d-49d3cd6b0958/nova-manage/0.log" Oct 02 17:03:21 crc kubenswrapper[4808]: I1002 17:03:21.952289 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_eb046ea2-150a-494e-bfa5-5b4283f62456/nova-cell0-conductor-conductor/0.log" Oct 02 17:03:22 crc kubenswrapper[4808]: I1002 17:03:22.171592 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-db-sync-hmw58_81cc1bfd-c3b7-49e1-9bc7-466483bde77b/nova-cell0-conductor-db-sync/0.log" Oct 02 17:03:22 crc kubenswrapper[4808]: I1002 17:03:22.381395 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-db-create-x5dj9_1c64446b-d171-4749-a4b6-0b71c80ce9da/mariadb-database-create/0.log" Oct 02 17:03:22 crc kubenswrapper[4808]: I1002 17:03:22.577793 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-4976-account-create-lskbv_b8311380-1e59-4d6c-b48f-c7dac382f325/mariadb-account-create/0.log" Oct 02 17:03:22 crc kubenswrapper[4808]: I1002 17:03:22.757490 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-cell-mapping-jkwct_802da733-8c7c-450b-a9c6-7fda3f2d2fbc/nova-manage/0.log" Oct 02 17:03:23 crc kubenswrapper[4808]: I1002 17:03:23.014789 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_75314254-401d-44e1-b68b-658aad79a9df/nova-cell1-conductor-conductor/0.log" Oct 02 17:03:23 crc kubenswrapper[4808]: I1002 17:03:23.206679 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-db-sync-z46bq_d4c1d967-acf9-4dbb-83db-1f654d4be6d2/nova-cell1-conductor-db-sync/0.log" Oct 02 17:03:23 crc kubenswrapper[4808]: I1002 17:03:23.394624 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-db-create-nkghg_7921b775-ed41-47c3-b334-0345e978f9d0/mariadb-database-create/0.log" Oct 02 17:03:23 crc kubenswrapper[4808]: I1002 17:03:23.606268 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_d89cc77e-ed29-4940-8db0-b7c32499d59e/nova-cell1-novncproxy-novncproxy/0.log" Oct 02 17:03:23 crc kubenswrapper[4808]: I1002 17:03:23.841868 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_c0748c83-a144-4c4c-9980-41c6592edf0f/nova-metadata-log/0.log" Oct 02 17:03:24 crc kubenswrapper[4808]: I1002 17:03:24.007424 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_c0748c83-a144-4c4c-9980-41c6592edf0f/nova-metadata-metadata/0.log" Oct 02 17:03:24 crc kubenswrapper[4808]: I1002 17:03:24.312331 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_1cebaac7-db1e-4020-ab4a-b63ebf95124d/nova-scheduler-scheduler/0.log" Oct 02 17:03:24 crc kubenswrapper[4808]: I1002 17:03:24.453718 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a8509820-f0d7-417b-92d4-1be8b7b714ef/mysql-bootstrap/0.log" Oct 02 17:03:24 crc kubenswrapper[4808]: I1002 17:03:24.648560 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a8509820-f0d7-417b-92d4-1be8b7b714ef/mysql-bootstrap/0.log" Oct 02 17:03:24 crc kubenswrapper[4808]: I1002 17:03:24.687466 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a8509820-f0d7-417b-92d4-1be8b7b714ef/galera/0.log" Oct 02 17:03:24 crc kubenswrapper[4808]: I1002 17:03:24.923001 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_500b4d21-bf0e-4763-91bd-7c32bf13a064/mysql-bootstrap/0.log" Oct 02 17:03:25 crc kubenswrapper[4808]: I1002 17:03:25.171718 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_500b4d21-bf0e-4763-91bd-7c32bf13a064/mysql-bootstrap/0.log" Oct 02 17:03:25 crc kubenswrapper[4808]: I1002 17:03:25.187483 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_500b4d21-bf0e-4763-91bd-7c32bf13a064/galera/0.log" Oct 02 17:03:25 crc kubenswrapper[4808]: I1002 17:03:25.379665 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_79b74b54-a9e1-4a7c-841a-1b4d4588b6bf/openstackclient/0.log" Oct 02 17:03:25 crc kubenswrapper[4808]: I1002 17:03:25.543123 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-hzbv4_eafc4ad2-d1ad-446a-ae4e-4fd250aaa5f8/ovn-controller/0.log" Oct 02 17:03:25 crc kubenswrapper[4808]: I1002 17:03:25.782100 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-8d9tc_ebecde28-66be-4caa-87d4-779d8605da87/openstack-network-exporter/0.log" Oct 02 17:03:25 crc kubenswrapper[4808]: I1002 17:03:25.891616 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_8b0cce8a-5f59-490d-87e6-6186890cd2a8/memcached/0.log" Oct 02 17:03:25 crc kubenswrapper[4808]: I1002 17:03:25.984660 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-59c8c_fd2a1d38-1169-4ab7-995f-eb6fb2f230a6/ovsdb-server-init/0.log" Oct 02 17:03:26 crc kubenswrapper[4808]: I1002 17:03:26.128414 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-59c8c_fd2a1d38-1169-4ab7-995f-eb6fb2f230a6/ovsdb-server-init/0.log" Oct 02 17:03:26 crc kubenswrapper[4808]: I1002 17:03:26.146499 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-59c8c_fd2a1d38-1169-4ab7-995f-eb6fb2f230a6/ovs-vswitchd/0.log" Oct 02 17:03:26 crc kubenswrapper[4808]: I1002 17:03:26.148688 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-59c8c_fd2a1d38-1169-4ab7-995f-eb6fb2f230a6/ovsdb-server/0.log" Oct 02 17:03:26 crc kubenswrapper[4808]: I1002 17:03:26.330625 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c84f87a5-f6b6-49b9-aa85-e88695deaa87/openstack-network-exporter/0.log" Oct 02 17:03:26 crc kubenswrapper[4808]: I1002 17:03:26.335334 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c84f87a5-f6b6-49b9-aa85-e88695deaa87/ovn-northd/0.log" Oct 02 17:03:26 crc kubenswrapper[4808]: I1002 17:03:26.482804 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d642175e-72ec-466b-9273-b82beb5e99f5/openstack-network-exporter/0.log" Oct 02 17:03:26 crc kubenswrapper[4808]: I1002 17:03:26.524510 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d642175e-72ec-466b-9273-b82beb5e99f5/ovsdbserver-nb/0.log" Oct 02 17:03:26 crc kubenswrapper[4808]: I1002 17:03:26.653514 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_dbb01d1d-5fc2-4d1c-8fa3-5575de879622/openstack-network-exporter/0.log" Oct 02 17:03:26 crc kubenswrapper[4808]: I1002 17:03:26.693837 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_dbb01d1d-5fc2-4d1c-8fa3-5575de879622/ovsdbserver-sb/0.log" Oct 02 17:03:26 crc kubenswrapper[4808]: I1002 17:03:26.845388 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7dbcfb7656-7hqb7_6317ce08-fda5-4271-908c-a18a0fd29180/placement-api/0.log" Oct 02 17:03:26 crc kubenswrapper[4808]: I1002 17:03:26.912104 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7dbcfb7656-7hqb7_6317ce08-fda5-4271-908c-a18a0fd29180/placement-log/0.log" Oct 02 17:03:27 crc kubenswrapper[4808]: I1002 17:03:27.026476 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-828c-account-create-6vmr9_bf2f1836-9826-4d1c-89ea-eb979de32e62/mariadb-account-create/0.log" Oct 02 17:03:27 crc kubenswrapper[4808]: I1002 17:03:27.197010 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-db-create-5fthx_0b6c4102-3ccd-4119-bbad-697ad2589e40/mariadb-database-create/0.log" Oct 02 17:03:27 crc kubenswrapper[4808]: I1002 17:03:27.259317 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-db-sync-dl6kd_7a437827-755a-4648-a40e-18a8ef89355d/placement-db-sync/0.log" Oct 02 17:03:27 crc kubenswrapper[4808]: I1002 17:03:27.425170 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_49331c3b-ac26-4f45-ad92-fadf45724c42/setup-container/0.log" Oct 02 17:03:27 crc kubenswrapper[4808]: I1002 17:03:27.587568 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_49331c3b-ac26-4f45-ad92-fadf45724c42/setup-container/0.log" Oct 02 17:03:27 crc kubenswrapper[4808]: I1002 17:03:27.591646 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_49331c3b-ac26-4f45-ad92-fadf45724c42/rabbitmq/0.log" Oct 02 17:03:27 crc kubenswrapper[4808]: I1002 17:03:27.759928 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_99aebb2e-33d8-4657-9ed2-cf23aeb8a281/setup-container/0.log" Oct 02 17:03:27 crc kubenswrapper[4808]: I1002 17:03:27.900576 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_99aebb2e-33d8-4657-9ed2-cf23aeb8a281/setup-container/0.log" Oct 02 17:03:27 crc kubenswrapper[4808]: I1002 17:03:27.986151 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_99aebb2e-33d8-4657-9ed2-cf23aeb8a281/rabbitmq/0.log" Oct 02 17:03:28 crc kubenswrapper[4808]: I1002 17:03:28.087163 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-gd897_c04a4cc7-6d1c-419e-9751-94b14f873801/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 17:03:34 crc kubenswrapper[4808]: I1002 17:03:34.830854 4808 scope.go:117] "RemoveContainer" containerID="df1bd60d449001a0deb4d9dbe2d011019d0d5f09a9a1de3b8bbd2e2c45e06b03" Oct 02 17:03:34 crc kubenswrapper[4808]: I1002 17:03:34.862031 4808 scope.go:117] "RemoveContainer" containerID="4418d941a635dd08c35aae363cd8749f5fc4adbad6b78fe5924c550b448290f0" Oct 02 17:03:45 crc kubenswrapper[4808]: I1002 17:03:45.306450 4808 generic.go:334] "Generic (PLEG): container finished" podID="3c51fb4d-7c38-449f-8d50-6bb8f34823b1" containerID="2986dd928bde6cf69270dddd5c0f22363ef81625c628a0bcbea2b8e1878e72bc" exitCode=0 Oct 02 17:03:45 crc kubenswrapper[4808]: I1002 17:03:45.306519 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jdsw/crc-debug-8w4dv" event={"ID":"3c51fb4d-7c38-449f-8d50-6bb8f34823b1","Type":"ContainerDied","Data":"2986dd928bde6cf69270dddd5c0f22363ef81625c628a0bcbea2b8e1878e72bc"} Oct 02 17:03:46 crc kubenswrapper[4808]: I1002 17:03:46.436771 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jdsw/crc-debug-8w4dv" Oct 02 17:03:46 crc kubenswrapper[4808]: I1002 17:03:46.463428 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5jdsw/crc-debug-8w4dv"] Oct 02 17:03:46 crc kubenswrapper[4808]: I1002 17:03:46.470974 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5jdsw/crc-debug-8w4dv"] Oct 02 17:03:46 crc kubenswrapper[4808]: I1002 17:03:46.590765 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3c51fb4d-7c38-449f-8d50-6bb8f34823b1-host\") pod \"3c51fb4d-7c38-449f-8d50-6bb8f34823b1\" (UID: \"3c51fb4d-7c38-449f-8d50-6bb8f34823b1\") " Oct 02 17:03:46 crc kubenswrapper[4808]: I1002 17:03:46.590924 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8cqq\" (UniqueName: \"kubernetes.io/projected/3c51fb4d-7c38-449f-8d50-6bb8f34823b1-kube-api-access-w8cqq\") pod \"3c51fb4d-7c38-449f-8d50-6bb8f34823b1\" (UID: \"3c51fb4d-7c38-449f-8d50-6bb8f34823b1\") " Oct 02 17:03:46 crc kubenswrapper[4808]: I1002 17:03:46.590916 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c51fb4d-7c38-449f-8d50-6bb8f34823b1-host" (OuterVolumeSpecName: "host") pod "3c51fb4d-7c38-449f-8d50-6bb8f34823b1" (UID: "3c51fb4d-7c38-449f-8d50-6bb8f34823b1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 17:03:46 crc kubenswrapper[4808]: I1002 17:03:46.591464 4808 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3c51fb4d-7c38-449f-8d50-6bb8f34823b1-host\") on node \"crc\" DevicePath \"\"" Oct 02 17:03:46 crc kubenswrapper[4808]: I1002 17:03:46.603425 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c51fb4d-7c38-449f-8d50-6bb8f34823b1-kube-api-access-w8cqq" (OuterVolumeSpecName: "kube-api-access-w8cqq") pod "3c51fb4d-7c38-449f-8d50-6bb8f34823b1" (UID: "3c51fb4d-7c38-449f-8d50-6bb8f34823b1"). InnerVolumeSpecName "kube-api-access-w8cqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:03:46 crc kubenswrapper[4808]: I1002 17:03:46.693408 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8cqq\" (UniqueName: \"kubernetes.io/projected/3c51fb4d-7c38-449f-8d50-6bb8f34823b1-kube-api-access-w8cqq\") on node \"crc\" DevicePath \"\"" Oct 02 17:03:47 crc kubenswrapper[4808]: I1002 17:03:47.326640 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91cd411289a8626c96347b5f2e2c20c8fac8e0b2c647b0c0c4924f5056271f3b" Oct 02 17:03:47 crc kubenswrapper[4808]: I1002 17:03:47.326703 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jdsw/crc-debug-8w4dv" Oct 02 17:03:47 crc kubenswrapper[4808]: I1002 17:03:47.414711 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c51fb4d-7c38-449f-8d50-6bb8f34823b1" path="/var/lib/kubelet/pods/3c51fb4d-7c38-449f-8d50-6bb8f34823b1/volumes" Oct 02 17:03:47 crc kubenswrapper[4808]: I1002 17:03:47.675016 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5jdsw/crc-debug-h878k"] Oct 02 17:03:47 crc kubenswrapper[4808]: E1002 17:03:47.675382 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c51fb4d-7c38-449f-8d50-6bb8f34823b1" containerName="container-00" Oct 02 17:03:47 crc kubenswrapper[4808]: I1002 17:03:47.675395 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c51fb4d-7c38-449f-8d50-6bb8f34823b1" containerName="container-00" Oct 02 17:03:47 crc kubenswrapper[4808]: I1002 17:03:47.675547 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c51fb4d-7c38-449f-8d50-6bb8f34823b1" containerName="container-00" Oct 02 17:03:47 crc kubenswrapper[4808]: I1002 17:03:47.676105 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jdsw/crc-debug-h878k" Oct 02 17:03:47 crc kubenswrapper[4808]: I1002 17:03:47.678391 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-5jdsw"/"default-dockercfg-nm2z4" Oct 02 17:03:47 crc kubenswrapper[4808]: I1002 17:03:47.813385 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f4a19a88-448b-458f-9c01-18c8eb2888b0-host\") pod \"crc-debug-h878k\" (UID: \"f4a19a88-448b-458f-9c01-18c8eb2888b0\") " pod="openshift-must-gather-5jdsw/crc-debug-h878k" Oct 02 17:03:47 crc kubenswrapper[4808]: I1002 17:03:47.813559 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c98s5\" (UniqueName: \"kubernetes.io/projected/f4a19a88-448b-458f-9c01-18c8eb2888b0-kube-api-access-c98s5\") pod \"crc-debug-h878k\" (UID: \"f4a19a88-448b-458f-9c01-18c8eb2888b0\") " pod="openshift-must-gather-5jdsw/crc-debug-h878k" Oct 02 17:03:47 crc kubenswrapper[4808]: I1002 17:03:47.914622 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f4a19a88-448b-458f-9c01-18c8eb2888b0-host\") pod \"crc-debug-h878k\" (UID: \"f4a19a88-448b-458f-9c01-18c8eb2888b0\") " pod="openshift-must-gather-5jdsw/crc-debug-h878k" Oct 02 17:03:47 crc kubenswrapper[4808]: I1002 17:03:47.914741 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f4a19a88-448b-458f-9c01-18c8eb2888b0-host\") pod \"crc-debug-h878k\" (UID: \"f4a19a88-448b-458f-9c01-18c8eb2888b0\") " pod="openshift-must-gather-5jdsw/crc-debug-h878k" Oct 02 17:03:47 crc kubenswrapper[4808]: I1002 17:03:47.914763 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c98s5\" (UniqueName: \"kubernetes.io/projected/f4a19a88-448b-458f-9c01-18c8eb2888b0-kube-api-access-c98s5\") pod \"crc-debug-h878k\" (UID: \"f4a19a88-448b-458f-9c01-18c8eb2888b0\") " pod="openshift-must-gather-5jdsw/crc-debug-h878k" Oct 02 17:03:47 crc kubenswrapper[4808]: I1002 17:03:47.935067 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c98s5\" (UniqueName: \"kubernetes.io/projected/f4a19a88-448b-458f-9c01-18c8eb2888b0-kube-api-access-c98s5\") pod \"crc-debug-h878k\" (UID: \"f4a19a88-448b-458f-9c01-18c8eb2888b0\") " pod="openshift-must-gather-5jdsw/crc-debug-h878k" Oct 02 17:03:47 crc kubenswrapper[4808]: I1002 17:03:47.989773 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jdsw/crc-debug-h878k" Oct 02 17:03:48 crc kubenswrapper[4808]: I1002 17:03:48.336629 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jdsw/crc-debug-h878k" event={"ID":"f4a19a88-448b-458f-9c01-18c8eb2888b0","Type":"ContainerStarted","Data":"1ade23720a7d651b49e634bb49c3db01a36dc269c97aba55612c65f505148807"} Oct 02 17:03:48 crc kubenswrapper[4808]: I1002 17:03:48.336963 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jdsw/crc-debug-h878k" event={"ID":"f4a19a88-448b-458f-9c01-18c8eb2888b0","Type":"ContainerStarted","Data":"cd82372d87edb552d5a2e571faee9e70105fecca8f183bcf0477c08b30e6c916"} Oct 02 17:03:48 crc kubenswrapper[4808]: I1002 17:03:48.348922 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-5jdsw/crc-debug-h878k" podStartSLOduration=1.3489059700000001 podStartE2EDuration="1.34890597s" podCreationTimestamp="2025-10-02 17:03:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 17:03:48.345355035 +0000 UTC m=+1395.670884035" watchObservedRunningTime="2025-10-02 17:03:48.34890597 +0000 UTC m=+1395.674434970" Oct 02 17:03:49 crc kubenswrapper[4808]: I1002 17:03:49.344201 4808 generic.go:334] "Generic (PLEG): container finished" podID="f4a19a88-448b-458f-9c01-18c8eb2888b0" containerID="1ade23720a7d651b49e634bb49c3db01a36dc269c97aba55612c65f505148807" exitCode=0 Oct 02 17:03:49 crc kubenswrapper[4808]: I1002 17:03:49.344468 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jdsw/crc-debug-h878k" event={"ID":"f4a19a88-448b-458f-9c01-18c8eb2888b0","Type":"ContainerDied","Data":"1ade23720a7d651b49e634bb49c3db01a36dc269c97aba55612c65f505148807"} Oct 02 17:03:50 crc kubenswrapper[4808]: I1002 17:03:50.442946 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jdsw/crc-debug-h878k" Oct 02 17:03:50 crc kubenswrapper[4808]: I1002 17:03:50.551675 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c98s5\" (UniqueName: \"kubernetes.io/projected/f4a19a88-448b-458f-9c01-18c8eb2888b0-kube-api-access-c98s5\") pod \"f4a19a88-448b-458f-9c01-18c8eb2888b0\" (UID: \"f4a19a88-448b-458f-9c01-18c8eb2888b0\") " Oct 02 17:03:50 crc kubenswrapper[4808]: I1002 17:03:50.551782 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f4a19a88-448b-458f-9c01-18c8eb2888b0-host\") pod \"f4a19a88-448b-458f-9c01-18c8eb2888b0\" (UID: \"f4a19a88-448b-458f-9c01-18c8eb2888b0\") " Oct 02 17:03:50 crc kubenswrapper[4808]: I1002 17:03:50.553351 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4a19a88-448b-458f-9c01-18c8eb2888b0-host" (OuterVolumeSpecName: "host") pod "f4a19a88-448b-458f-9c01-18c8eb2888b0" (UID: "f4a19a88-448b-458f-9c01-18c8eb2888b0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 17:03:50 crc kubenswrapper[4808]: I1002 17:03:50.565771 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4a19a88-448b-458f-9c01-18c8eb2888b0-kube-api-access-c98s5" (OuterVolumeSpecName: "kube-api-access-c98s5") pod "f4a19a88-448b-458f-9c01-18c8eb2888b0" (UID: "f4a19a88-448b-458f-9c01-18c8eb2888b0"). InnerVolumeSpecName "kube-api-access-c98s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:03:50 crc kubenswrapper[4808]: I1002 17:03:50.653322 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c98s5\" (UniqueName: \"kubernetes.io/projected/f4a19a88-448b-458f-9c01-18c8eb2888b0-kube-api-access-c98s5\") on node \"crc\" DevicePath \"\"" Oct 02 17:03:50 crc kubenswrapper[4808]: I1002 17:03:50.653347 4808 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f4a19a88-448b-458f-9c01-18c8eb2888b0-host\") on node \"crc\" DevicePath \"\"" Oct 02 17:03:51 crc kubenswrapper[4808]: I1002 17:03:51.364534 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jdsw/crc-debug-h878k" event={"ID":"f4a19a88-448b-458f-9c01-18c8eb2888b0","Type":"ContainerDied","Data":"cd82372d87edb552d5a2e571faee9e70105fecca8f183bcf0477c08b30e6c916"} Oct 02 17:03:51 crc kubenswrapper[4808]: I1002 17:03:51.364568 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd82372d87edb552d5a2e571faee9e70105fecca8f183bcf0477c08b30e6c916" Oct 02 17:03:51 crc kubenswrapper[4808]: I1002 17:03:51.364584 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jdsw/crc-debug-h878k" Oct 02 17:03:52 crc kubenswrapper[4808]: I1002 17:03:52.549421 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5jdsw/crc-debug-h878k"] Oct 02 17:03:52 crc kubenswrapper[4808]: I1002 17:03:52.557319 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5jdsw/crc-debug-h878k"] Oct 02 17:03:53 crc kubenswrapper[4808]: I1002 17:03:53.410901 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4a19a88-448b-458f-9c01-18c8eb2888b0" path="/var/lib/kubelet/pods/f4a19a88-448b-458f-9c01-18c8eb2888b0/volumes" Oct 02 17:03:53 crc kubenswrapper[4808]: I1002 17:03:53.806864 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5jdsw/crc-debug-b7pbk"] Oct 02 17:03:53 crc kubenswrapper[4808]: E1002 17:03:53.808593 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4a19a88-448b-458f-9c01-18c8eb2888b0" containerName="container-00" Oct 02 17:03:53 crc kubenswrapper[4808]: I1002 17:03:53.808821 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4a19a88-448b-458f-9c01-18c8eb2888b0" containerName="container-00" Oct 02 17:03:53 crc kubenswrapper[4808]: I1002 17:03:53.809352 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4a19a88-448b-458f-9c01-18c8eb2888b0" containerName="container-00" Oct 02 17:03:53 crc kubenswrapper[4808]: I1002 17:03:53.810558 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jdsw/crc-debug-b7pbk" Oct 02 17:03:53 crc kubenswrapper[4808]: I1002 17:03:53.815571 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-5jdsw"/"default-dockercfg-nm2z4" Oct 02 17:03:53 crc kubenswrapper[4808]: I1002 17:03:53.907099 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/49527641-174a-40bf-814f-ad9f90470eb0-host\") pod \"crc-debug-b7pbk\" (UID: \"49527641-174a-40bf-814f-ad9f90470eb0\") " pod="openshift-must-gather-5jdsw/crc-debug-b7pbk" Oct 02 17:03:53 crc kubenswrapper[4808]: I1002 17:03:53.907339 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl7dc\" (UniqueName: \"kubernetes.io/projected/49527641-174a-40bf-814f-ad9f90470eb0-kube-api-access-vl7dc\") pod \"crc-debug-b7pbk\" (UID: \"49527641-174a-40bf-814f-ad9f90470eb0\") " pod="openshift-must-gather-5jdsw/crc-debug-b7pbk" Oct 02 17:03:54 crc kubenswrapper[4808]: I1002 17:03:54.008550 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/49527641-174a-40bf-814f-ad9f90470eb0-host\") pod \"crc-debug-b7pbk\" (UID: \"49527641-174a-40bf-814f-ad9f90470eb0\") " pod="openshift-must-gather-5jdsw/crc-debug-b7pbk" Oct 02 17:03:54 crc kubenswrapper[4808]: I1002 17:03:54.008704 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl7dc\" (UniqueName: \"kubernetes.io/projected/49527641-174a-40bf-814f-ad9f90470eb0-kube-api-access-vl7dc\") pod \"crc-debug-b7pbk\" (UID: \"49527641-174a-40bf-814f-ad9f90470eb0\") " pod="openshift-must-gather-5jdsw/crc-debug-b7pbk" Oct 02 17:03:54 crc kubenswrapper[4808]: I1002 17:03:54.008969 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/49527641-174a-40bf-814f-ad9f90470eb0-host\") pod \"crc-debug-b7pbk\" (UID: \"49527641-174a-40bf-814f-ad9f90470eb0\") " pod="openshift-must-gather-5jdsw/crc-debug-b7pbk" Oct 02 17:03:54 crc kubenswrapper[4808]: I1002 17:03:54.032669 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl7dc\" (UniqueName: \"kubernetes.io/projected/49527641-174a-40bf-814f-ad9f90470eb0-kube-api-access-vl7dc\") pod \"crc-debug-b7pbk\" (UID: \"49527641-174a-40bf-814f-ad9f90470eb0\") " pod="openshift-must-gather-5jdsw/crc-debug-b7pbk" Oct 02 17:03:54 crc kubenswrapper[4808]: I1002 17:03:54.131800 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jdsw/crc-debug-b7pbk" Oct 02 17:03:54 crc kubenswrapper[4808]: I1002 17:03:54.400131 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jdsw/crc-debug-b7pbk" event={"ID":"49527641-174a-40bf-814f-ad9f90470eb0","Type":"ContainerStarted","Data":"51b0bdf66a935343af8e167f946403ef9e5cf578a21403bc4fd41f08f3f23bfd"} Oct 02 17:03:55 crc kubenswrapper[4808]: I1002 17:03:55.421476 4808 generic.go:334] "Generic (PLEG): container finished" podID="49527641-174a-40bf-814f-ad9f90470eb0" containerID="29b6016a2f570a152d7c01884c9b4d5f1d532cb8606aa80602b6e5e1a8757706" exitCode=0 Oct 02 17:03:55 crc kubenswrapper[4808]: I1002 17:03:55.422166 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jdsw/crc-debug-b7pbk" event={"ID":"49527641-174a-40bf-814f-ad9f90470eb0","Type":"ContainerDied","Data":"29b6016a2f570a152d7c01884c9b4d5f1d532cb8606aa80602b6e5e1a8757706"} Oct 02 17:03:55 crc kubenswrapper[4808]: I1002 17:03:55.464230 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5jdsw/crc-debug-b7pbk"] Oct 02 17:03:55 crc kubenswrapper[4808]: I1002 17:03:55.474257 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5jdsw/crc-debug-b7pbk"] Oct 02 17:03:56 crc kubenswrapper[4808]: I1002 17:03:56.549154 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jdsw/crc-debug-b7pbk" Oct 02 17:03:56 crc kubenswrapper[4808]: I1002 17:03:56.665033 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/49527641-174a-40bf-814f-ad9f90470eb0-host\") pod \"49527641-174a-40bf-814f-ad9f90470eb0\" (UID: \"49527641-174a-40bf-814f-ad9f90470eb0\") " Oct 02 17:03:56 crc kubenswrapper[4808]: I1002 17:03:56.665103 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vl7dc\" (UniqueName: \"kubernetes.io/projected/49527641-174a-40bf-814f-ad9f90470eb0-kube-api-access-vl7dc\") pod \"49527641-174a-40bf-814f-ad9f90470eb0\" (UID: \"49527641-174a-40bf-814f-ad9f90470eb0\") " Oct 02 17:03:56 crc kubenswrapper[4808]: I1002 17:03:56.665173 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/49527641-174a-40bf-814f-ad9f90470eb0-host" (OuterVolumeSpecName: "host") pod "49527641-174a-40bf-814f-ad9f90470eb0" (UID: "49527641-174a-40bf-814f-ad9f90470eb0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 17:03:56 crc kubenswrapper[4808]: I1002 17:03:56.665450 4808 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/49527641-174a-40bf-814f-ad9f90470eb0-host\") on node \"crc\" DevicePath \"\"" Oct 02 17:03:56 crc kubenswrapper[4808]: I1002 17:03:56.680217 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49527641-174a-40bf-814f-ad9f90470eb0-kube-api-access-vl7dc" (OuterVolumeSpecName: "kube-api-access-vl7dc") pod "49527641-174a-40bf-814f-ad9f90470eb0" (UID: "49527641-174a-40bf-814f-ad9f90470eb0"). InnerVolumeSpecName "kube-api-access-vl7dc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:03:56 crc kubenswrapper[4808]: I1002 17:03:56.767121 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vl7dc\" (UniqueName: \"kubernetes.io/projected/49527641-174a-40bf-814f-ad9f90470eb0-kube-api-access-vl7dc\") on node \"crc\" DevicePath \"\"" Oct 02 17:03:57 crc kubenswrapper[4808]: I1002 17:03:57.303256 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf_8771d707-d0d8-4b3b-8ca5-d68949b29b64/util/0.log" Oct 02 17:03:57 crc kubenswrapper[4808]: I1002 17:03:57.406721 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49527641-174a-40bf-814f-ad9f90470eb0" path="/var/lib/kubelet/pods/49527641-174a-40bf-814f-ad9f90470eb0/volumes" Oct 02 17:03:57 crc kubenswrapper[4808]: I1002 17:03:57.432611 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf_8771d707-d0d8-4b3b-8ca5-d68949b29b64/util/0.log" Oct 02 17:03:57 crc kubenswrapper[4808]: I1002 17:03:57.435064 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf_8771d707-d0d8-4b3b-8ca5-d68949b29b64/pull/0.log" Oct 02 17:03:57 crc kubenswrapper[4808]: I1002 17:03:57.437342 4808 scope.go:117] "RemoveContainer" containerID="29b6016a2f570a152d7c01884c9b4d5f1d532cb8606aa80602b6e5e1a8757706" Oct 02 17:03:57 crc kubenswrapper[4808]: I1002 17:03:57.437355 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jdsw/crc-debug-b7pbk" Oct 02 17:03:57 crc kubenswrapper[4808]: I1002 17:03:57.513300 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf_8771d707-d0d8-4b3b-8ca5-d68949b29b64/pull/0.log" Oct 02 17:03:57 crc kubenswrapper[4808]: I1002 17:03:57.669767 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf_8771d707-d0d8-4b3b-8ca5-d68949b29b64/extract/0.log" Oct 02 17:03:57 crc kubenswrapper[4808]: I1002 17:03:57.698709 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf_8771d707-d0d8-4b3b-8ca5-d68949b29b64/util/0.log" Oct 02 17:03:57 crc kubenswrapper[4808]: I1002 17:03:57.712462 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca698aa2253c044890ca9ff1a3dc138e56ab71add637faf800fa325dabhtsf_8771d707-d0d8-4b3b-8ca5-d68949b29b64/pull/0.log" Oct 02 17:03:57 crc kubenswrapper[4808]: I1002 17:03:57.843618 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-mnftb_6253eb77-9a32-473f-8f30-837d648e51cb/kube-rbac-proxy/0.log" Oct 02 17:03:57 crc kubenswrapper[4808]: I1002 17:03:57.875327 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-gnv8z_e1981fc0-4eab-4cec-8c2a-611ab0d2af38/kube-rbac-proxy/0.log" Oct 02 17:03:57 crc kubenswrapper[4808]: I1002 17:03:57.917956 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-mnftb_6253eb77-9a32-473f-8f30-837d648e51cb/manager/0.log" Oct 02 17:03:58 crc kubenswrapper[4808]: I1002 17:03:58.030159 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-gnv8z_e1981fc0-4eab-4cec-8c2a-611ab0d2af38/manager/0.log" Oct 02 17:03:58 crc kubenswrapper[4808]: I1002 17:03:58.101960 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-2zhp9_7546f33a-e8d7-4e15-bb6f-bd91a95d7ae9/kube-rbac-proxy/0.log" Oct 02 17:03:58 crc kubenswrapper[4808]: I1002 17:03:58.188928 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-2zhp9_7546f33a-e8d7-4e15-bb6f-bd91a95d7ae9/manager/0.log" Oct 02 17:03:58 crc kubenswrapper[4808]: I1002 17:03:58.290577 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b99c7c5b-z2gb5_0193c2d7-ccc0-40bc-b4a9-e483e51803ce/kube-rbac-proxy/0.log" Oct 02 17:03:58 crc kubenswrapper[4808]: I1002 17:03:58.369000 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b99c7c5b-z2gb5_0193c2d7-ccc0-40bc-b4a9-e483e51803ce/manager/0.log" Oct 02 17:03:58 crc kubenswrapper[4808]: I1002 17:03:58.412021 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-m5tw2_8771edaf-8b38-4efe-9a99-ea74be174f75/kube-rbac-proxy/0.log" Oct 02 17:03:58 crc kubenswrapper[4808]: I1002 17:03:58.486257 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-m5tw2_8771edaf-8b38-4efe-9a99-ea74be174f75/manager/0.log" Oct 02 17:03:58 crc kubenswrapper[4808]: I1002 17:03:58.573198 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-6qnjp_4a8f2a74-9110-4f79-bc84-c68ed74b982d/kube-rbac-proxy/0.log" Oct 02 17:03:58 crc kubenswrapper[4808]: I1002 17:03:58.600953 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-6qnjp_4a8f2a74-9110-4f79-bc84-c68ed74b982d/manager/0.log" Oct 02 17:03:58 crc kubenswrapper[4808]: I1002 17:03:58.697259 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-kjrph_e2c80909-0584-4f0c-a989-fe053b196f98/kube-rbac-proxy/0.log" Oct 02 17:03:58 crc kubenswrapper[4808]: I1002 17:03:58.835080 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-9cm78_4b8b017f-fc2b-4925-ab5c-001ca052ff8c/kube-rbac-proxy/0.log" Oct 02 17:03:58 crc kubenswrapper[4808]: I1002 17:03:58.874305 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-kjrph_e2c80909-0584-4f0c-a989-fe053b196f98/manager/0.log" Oct 02 17:03:58 crc kubenswrapper[4808]: I1002 17:03:58.900499 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-9cm78_4b8b017f-fc2b-4925-ab5c-001ca052ff8c/manager/0.log" Oct 02 17:03:59 crc kubenswrapper[4808]: I1002 17:03:59.013961 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-65f4bcbb94-svkdd_818e02a3-653a-4b36-a68b-9d4e2df0b71e/kube-rbac-proxy/0.log" Oct 02 17:03:59 crc kubenswrapper[4808]: I1002 17:03:59.065221 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-65f4bcbb94-svkdd_818e02a3-653a-4b36-a68b-9d4e2df0b71e/manager/3.log" Oct 02 17:03:59 crc kubenswrapper[4808]: I1002 17:03:59.099596 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-65f4bcbb94-svkdd_818e02a3-653a-4b36-a68b-9d4e2df0b71e/manager/2.log" Oct 02 17:03:59 crc kubenswrapper[4808]: I1002 17:03:59.205276 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-wpps5_14931328-29f9-4a03-9a3b-50cb2bb861fe/kube-rbac-proxy/0.log" Oct 02 17:03:59 crc kubenswrapper[4808]: I1002 17:03:59.250662 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-wpps5_14931328-29f9-4a03-9a3b-50cb2bb861fe/manager/0.log" Oct 02 17:03:59 crc kubenswrapper[4808]: I1002 17:03:59.413120 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-dqz59_43dd50ad-2f9d-47a6-b685-379dd593449c/kube-rbac-proxy/0.log" Oct 02 17:03:59 crc kubenswrapper[4808]: I1002 17:03:59.470135 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-dqz59_43dd50ad-2f9d-47a6-b685-379dd593449c/manager/0.log" Oct 02 17:03:59 crc kubenswrapper[4808]: I1002 17:03:59.495672 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-lz5ls_54002291-52de-4cd8-9c01-7a3bd3c2b98b/kube-rbac-proxy/0.log" Oct 02 17:03:59 crc kubenswrapper[4808]: I1002 17:03:59.619880 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-lz5ls_54002291-52de-4cd8-9c01-7a3bd3c2b98b/manager/0.log" Oct 02 17:03:59 crc kubenswrapper[4808]: I1002 17:03:59.638193 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-hsmt5_23f6eba8-a786-4234-869f-6e44016536dd/kube-rbac-proxy/0.log" Oct 02 17:03:59 crc kubenswrapper[4808]: I1002 17:03:59.750643 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-hsmt5_23f6eba8-a786-4234-869f-6e44016536dd/manager/0.log" Oct 02 17:03:59 crc kubenswrapper[4808]: I1002 17:03:59.816259 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-5tt4q_7737381d-608c-43f8-9cdf-a5612d28bd43/kube-rbac-proxy/0.log" Oct 02 17:03:59 crc kubenswrapper[4808]: I1002 17:03:59.900474 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-5tt4q_7737381d-608c-43f8-9cdf-a5612d28bd43/manager/0.log" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.032241 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5869cb545-zw4d5_56b8aff7-4d44-41c4-b4a7-caed261ab974/kube-rbac-proxy/0.log" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.054718 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5869cb545-zw4d5_56b8aff7-4d44-41c4-b4a7-caed261ab974/manager/0.log" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.153592 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6b57c44d8-px47q_4b8037b3-28b8-4492-b4a7-7ed5413df620/kube-rbac-proxy/0.log" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.300740 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fjtrs"] Oct 02 17:04:00 crc kubenswrapper[4808]: E1002 17:04:00.301111 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49527641-174a-40bf-814f-ad9f90470eb0" containerName="container-00" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.301128 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="49527641-174a-40bf-814f-ad9f90470eb0" containerName="container-00" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.301475 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="49527641-174a-40bf-814f-ad9f90470eb0" containerName="container-00" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.302664 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fjtrs" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.309177 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-755876968d-m4vm4_b0800622-881c-4e2a-9cb1-0447445eddc1/kube-rbac-proxy/0.log" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.328665 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fjtrs"] Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.429025 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/843834ff-31a7-4526-95bd-c37bff75cd13-utilities\") pod \"redhat-operators-fjtrs\" (UID: \"843834ff-31a7-4526-95bd-c37bff75cd13\") " pod="openshift-marketplace/redhat-operators-fjtrs" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.429254 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s74mp\" (UniqueName: \"kubernetes.io/projected/843834ff-31a7-4526-95bd-c37bff75cd13-kube-api-access-s74mp\") pod \"redhat-operators-fjtrs\" (UID: \"843834ff-31a7-4526-95bd-c37bff75cd13\") " pod="openshift-marketplace/redhat-operators-fjtrs" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.429668 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/843834ff-31a7-4526-95bd-c37bff75cd13-catalog-content\") pod \"redhat-operators-fjtrs\" (UID: \"843834ff-31a7-4526-95bd-c37bff75cd13\") " pod="openshift-marketplace/redhat-operators-fjtrs" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.496201 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zstrp_1883774e-2d09-400d-95bd-490deded815b/registry-server/0.log" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.531656 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/843834ff-31a7-4526-95bd-c37bff75cd13-catalog-content\") pod \"redhat-operators-fjtrs\" (UID: \"843834ff-31a7-4526-95bd-c37bff75cd13\") " pod="openshift-marketplace/redhat-operators-fjtrs" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.531787 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/843834ff-31a7-4526-95bd-c37bff75cd13-utilities\") pod \"redhat-operators-fjtrs\" (UID: \"843834ff-31a7-4526-95bd-c37bff75cd13\") " pod="openshift-marketplace/redhat-operators-fjtrs" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.531838 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s74mp\" (UniqueName: \"kubernetes.io/projected/843834ff-31a7-4526-95bd-c37bff75cd13-kube-api-access-s74mp\") pod \"redhat-operators-fjtrs\" (UID: \"843834ff-31a7-4526-95bd-c37bff75cd13\") " pod="openshift-marketplace/redhat-operators-fjtrs" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.532454 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/843834ff-31a7-4526-95bd-c37bff75cd13-catalog-content\") pod \"redhat-operators-fjtrs\" (UID: \"843834ff-31a7-4526-95bd-c37bff75cd13\") " pod="openshift-marketplace/redhat-operators-fjtrs" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.532776 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/843834ff-31a7-4526-95bd-c37bff75cd13-utilities\") pod \"redhat-operators-fjtrs\" (UID: \"843834ff-31a7-4526-95bd-c37bff75cd13\") " pod="openshift-marketplace/redhat-operators-fjtrs" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.553311 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s74mp\" (UniqueName: \"kubernetes.io/projected/843834ff-31a7-4526-95bd-c37bff75cd13-kube-api-access-s74mp\") pod \"redhat-operators-fjtrs\" (UID: \"843834ff-31a7-4526-95bd-c37bff75cd13\") " pod="openshift-marketplace/redhat-operators-fjtrs" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.559465 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-755876968d-m4vm4_b0800622-881c-4e2a-9cb1-0447445eddc1/operator/0.log" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.638216 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fjtrs" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.708408 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-hw4dk_0e51874f-4bce-4f5c-8e00-99f74b1f2881/kube-rbac-proxy/0.log" Oct 02 17:04:00 crc kubenswrapper[4808]: I1002 17:04:00.959525 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-g7gmf_0cf6cc81-9904-4904-bf45-a4d3e800231b/kube-rbac-proxy/0.log" Oct 02 17:04:01 crc kubenswrapper[4808]: I1002 17:04:01.002004 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6b57c44d8-px47q_4b8037b3-28b8-4492-b4a7-7ed5413df620/manager/0.log" Oct 02 17:04:01 crc kubenswrapper[4808]: I1002 17:04:01.102630 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fjtrs"] Oct 02 17:04:01 crc kubenswrapper[4808]: I1002 17:04:01.107038 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-hw4dk_0e51874f-4bce-4f5c-8e00-99f74b1f2881/manager/0.log" Oct 02 17:04:01 crc kubenswrapper[4808]: I1002 17:04:01.110542 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-g7gmf_0cf6cc81-9904-4904-bf45-a4d3e800231b/manager/0.log" Oct 02 17:04:01 crc kubenswrapper[4808]: I1002 17:04:01.312977 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-wv9sn_ccfa7a5b-15b3-4fd1-9a36-f04d894a838f/operator/0.log" Oct 02 17:04:01 crc kubenswrapper[4808]: I1002 17:04:01.377885 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-8bbm9_2a264f04-66b8-4938-ab3f-58c651166227/kube-rbac-proxy/0.log" Oct 02 17:04:01 crc kubenswrapper[4808]: I1002 17:04:01.428742 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-8bbm9_2a264f04-66b8-4938-ab3f-58c651166227/manager/0.log" Oct 02 17:04:01 crc kubenswrapper[4808]: I1002 17:04:01.468763 4808 generic.go:334] "Generic (PLEG): container finished" podID="843834ff-31a7-4526-95bd-c37bff75cd13" containerID="e714f42f509ea979f837ecd4e70b3cf082c7f64accd50387b1d79e7317f63a30" exitCode=0 Oct 02 17:04:01 crc kubenswrapper[4808]: I1002 17:04:01.468810 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjtrs" event={"ID":"843834ff-31a7-4526-95bd-c37bff75cd13","Type":"ContainerDied","Data":"e714f42f509ea979f837ecd4e70b3cf082c7f64accd50387b1d79e7317f63a30"} Oct 02 17:04:01 crc kubenswrapper[4808]: I1002 17:04:01.468855 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjtrs" event={"ID":"843834ff-31a7-4526-95bd-c37bff75cd13","Type":"ContainerStarted","Data":"f72ba19892b1f369a61fb96b957835a79fa2b84875349690d7cce7cf0119f6ad"} Oct 02 17:04:01 crc kubenswrapper[4808]: I1002 17:04:01.534904 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-vth7d_8b68ec2c-6583-482b-a9f4-d75fab859c95/kube-rbac-proxy/0.log" Oct 02 17:04:01 crc kubenswrapper[4808]: I1002 17:04:01.729719 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-vth7d_8b68ec2c-6583-482b-a9f4-d75fab859c95/manager/0.log" Oct 02 17:04:01 crc kubenswrapper[4808]: I1002 17:04:01.778216 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-skv6k_1ec1bdf1-04ec-4b70-8e58-5df6541b021f/kube-rbac-proxy/0.log" Oct 02 17:04:01 crc kubenswrapper[4808]: I1002 17:04:01.839940 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-9t5j7_d2828482-d7e8-4e5f-985d-197d614d8d15/kube-rbac-proxy/0.log" Oct 02 17:04:01 crc kubenswrapper[4808]: I1002 17:04:01.902459 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-skv6k_1ec1bdf1-04ec-4b70-8e58-5df6541b021f/manager/0.log" Oct 02 17:04:01 crc kubenswrapper[4808]: I1002 17:04:01.989490 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-9t5j7_d2828482-d7e8-4e5f-985d-197d614d8d15/manager/0.log" Oct 02 17:04:02 crc kubenswrapper[4808]: I1002 17:04:02.478195 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjtrs" event={"ID":"843834ff-31a7-4526-95bd-c37bff75cd13","Type":"ContainerStarted","Data":"feac121b0103962dcad8ef747f38ea06f2c27a818430a68817e88ffb3e8b15c6"} Oct 02 17:04:03 crc kubenswrapper[4808]: I1002 17:04:03.497209 4808 generic.go:334] "Generic (PLEG): container finished" podID="843834ff-31a7-4526-95bd-c37bff75cd13" containerID="feac121b0103962dcad8ef747f38ea06f2c27a818430a68817e88ffb3e8b15c6" exitCode=0 Oct 02 17:04:03 crc kubenswrapper[4808]: I1002 17:04:03.497400 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjtrs" event={"ID":"843834ff-31a7-4526-95bd-c37bff75cd13","Type":"ContainerDied","Data":"feac121b0103962dcad8ef747f38ea06f2c27a818430a68817e88ffb3e8b15c6"} Oct 02 17:04:04 crc kubenswrapper[4808]: I1002 17:04:04.506963 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjtrs" event={"ID":"843834ff-31a7-4526-95bd-c37bff75cd13","Type":"ContainerStarted","Data":"221e7e8e11106c9377288ae75a417ba4c07bcf60415d4db2cff6377df7bc440a"} Oct 02 17:04:04 crc kubenswrapper[4808]: I1002 17:04:04.527382 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fjtrs" podStartSLOduration=2.033494549 podStartE2EDuration="4.527366211s" podCreationTimestamp="2025-10-02 17:04:00 +0000 UTC" firstStartedPulling="2025-10-02 17:04:01.470381838 +0000 UTC m=+1408.795910838" lastFinishedPulling="2025-10-02 17:04:03.9642535 +0000 UTC m=+1411.289782500" observedRunningTime="2025-10-02 17:04:04.522339837 +0000 UTC m=+1411.847868837" watchObservedRunningTime="2025-10-02 17:04:04.527366211 +0000 UTC m=+1411.852895211" Oct 02 17:04:10 crc kubenswrapper[4808]: I1002 17:04:10.639682 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fjtrs" Oct 02 17:04:10 crc kubenswrapper[4808]: I1002 17:04:10.640338 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fjtrs" Oct 02 17:04:10 crc kubenswrapper[4808]: I1002 17:04:10.686960 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fjtrs" Oct 02 17:04:11 crc kubenswrapper[4808]: I1002 17:04:11.638314 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fjtrs" Oct 02 17:04:11 crc kubenswrapper[4808]: I1002 17:04:11.690269 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fjtrs"] Oct 02 17:04:13 crc kubenswrapper[4808]: I1002 17:04:13.584351 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fjtrs" podUID="843834ff-31a7-4526-95bd-c37bff75cd13" containerName="registry-server" containerID="cri-o://221e7e8e11106c9377288ae75a417ba4c07bcf60415d4db2cff6377df7bc440a" gracePeriod=2 Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.071198 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fjtrs" Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.186895 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/843834ff-31a7-4526-95bd-c37bff75cd13-catalog-content\") pod \"843834ff-31a7-4526-95bd-c37bff75cd13\" (UID: \"843834ff-31a7-4526-95bd-c37bff75cd13\") " Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.186964 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/843834ff-31a7-4526-95bd-c37bff75cd13-utilities\") pod \"843834ff-31a7-4526-95bd-c37bff75cd13\" (UID: \"843834ff-31a7-4526-95bd-c37bff75cd13\") " Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.186998 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s74mp\" (UniqueName: \"kubernetes.io/projected/843834ff-31a7-4526-95bd-c37bff75cd13-kube-api-access-s74mp\") pod \"843834ff-31a7-4526-95bd-c37bff75cd13\" (UID: \"843834ff-31a7-4526-95bd-c37bff75cd13\") " Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.187892 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/843834ff-31a7-4526-95bd-c37bff75cd13-utilities" (OuterVolumeSpecName: "utilities") pod "843834ff-31a7-4526-95bd-c37bff75cd13" (UID: "843834ff-31a7-4526-95bd-c37bff75cd13"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.195161 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/843834ff-31a7-4526-95bd-c37bff75cd13-kube-api-access-s74mp" (OuterVolumeSpecName: "kube-api-access-s74mp") pod "843834ff-31a7-4526-95bd-c37bff75cd13" (UID: "843834ff-31a7-4526-95bd-c37bff75cd13"). InnerVolumeSpecName "kube-api-access-s74mp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.281739 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/843834ff-31a7-4526-95bd-c37bff75cd13-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "843834ff-31a7-4526-95bd-c37bff75cd13" (UID: "843834ff-31a7-4526-95bd-c37bff75cd13"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.288936 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/843834ff-31a7-4526-95bd-c37bff75cd13-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.288961 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/843834ff-31a7-4526-95bd-c37bff75cd13-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.288971 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s74mp\" (UniqueName: \"kubernetes.io/projected/843834ff-31a7-4526-95bd-c37bff75cd13-kube-api-access-s74mp\") on node \"crc\" DevicePath \"\"" Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.596535 4808 generic.go:334] "Generic (PLEG): container finished" podID="843834ff-31a7-4526-95bd-c37bff75cd13" containerID="221e7e8e11106c9377288ae75a417ba4c07bcf60415d4db2cff6377df7bc440a" exitCode=0 Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.596579 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjtrs" event={"ID":"843834ff-31a7-4526-95bd-c37bff75cd13","Type":"ContainerDied","Data":"221e7e8e11106c9377288ae75a417ba4c07bcf60415d4db2cff6377df7bc440a"} Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.596613 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjtrs" event={"ID":"843834ff-31a7-4526-95bd-c37bff75cd13","Type":"ContainerDied","Data":"f72ba19892b1f369a61fb96b957835a79fa2b84875349690d7cce7cf0119f6ad"} Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.596630 4808 scope.go:117] "RemoveContainer" containerID="221e7e8e11106c9377288ae75a417ba4c07bcf60415d4db2cff6377df7bc440a" Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.596639 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fjtrs" Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.621087 4808 scope.go:117] "RemoveContainer" containerID="feac121b0103962dcad8ef747f38ea06f2c27a818430a68817e88ffb3e8b15c6" Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.644884 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fjtrs"] Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.653776 4808 scope.go:117] "RemoveContainer" containerID="e714f42f509ea979f837ecd4e70b3cf082c7f64accd50387b1d79e7317f63a30" Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.660004 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fjtrs"] Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.712883 4808 scope.go:117] "RemoveContainer" containerID="221e7e8e11106c9377288ae75a417ba4c07bcf60415d4db2cff6377df7bc440a" Oct 02 17:04:14 crc kubenswrapper[4808]: E1002 17:04:14.713314 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"221e7e8e11106c9377288ae75a417ba4c07bcf60415d4db2cff6377df7bc440a\": container with ID starting with 221e7e8e11106c9377288ae75a417ba4c07bcf60415d4db2cff6377df7bc440a not found: ID does not exist" containerID="221e7e8e11106c9377288ae75a417ba4c07bcf60415d4db2cff6377df7bc440a" Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.713348 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"221e7e8e11106c9377288ae75a417ba4c07bcf60415d4db2cff6377df7bc440a"} err="failed to get container status \"221e7e8e11106c9377288ae75a417ba4c07bcf60415d4db2cff6377df7bc440a\": rpc error: code = NotFound desc = could not find container \"221e7e8e11106c9377288ae75a417ba4c07bcf60415d4db2cff6377df7bc440a\": container with ID starting with 221e7e8e11106c9377288ae75a417ba4c07bcf60415d4db2cff6377df7bc440a not found: ID does not exist" Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.713368 4808 scope.go:117] "RemoveContainer" containerID="feac121b0103962dcad8ef747f38ea06f2c27a818430a68817e88ffb3e8b15c6" Oct 02 17:04:14 crc kubenswrapper[4808]: E1002 17:04:14.713681 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"feac121b0103962dcad8ef747f38ea06f2c27a818430a68817e88ffb3e8b15c6\": container with ID starting with feac121b0103962dcad8ef747f38ea06f2c27a818430a68817e88ffb3e8b15c6 not found: ID does not exist" containerID="feac121b0103962dcad8ef747f38ea06f2c27a818430a68817e88ffb3e8b15c6" Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.713702 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"feac121b0103962dcad8ef747f38ea06f2c27a818430a68817e88ffb3e8b15c6"} err="failed to get container status \"feac121b0103962dcad8ef747f38ea06f2c27a818430a68817e88ffb3e8b15c6\": rpc error: code = NotFound desc = could not find container \"feac121b0103962dcad8ef747f38ea06f2c27a818430a68817e88ffb3e8b15c6\": container with ID starting with feac121b0103962dcad8ef747f38ea06f2c27a818430a68817e88ffb3e8b15c6 not found: ID does not exist" Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.713714 4808 scope.go:117] "RemoveContainer" containerID="e714f42f509ea979f837ecd4e70b3cf082c7f64accd50387b1d79e7317f63a30" Oct 02 17:04:14 crc kubenswrapper[4808]: E1002 17:04:14.714051 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e714f42f509ea979f837ecd4e70b3cf082c7f64accd50387b1d79e7317f63a30\": container with ID starting with e714f42f509ea979f837ecd4e70b3cf082c7f64accd50387b1d79e7317f63a30 not found: ID does not exist" containerID="e714f42f509ea979f837ecd4e70b3cf082c7f64accd50387b1d79e7317f63a30" Oct 02 17:04:14 crc kubenswrapper[4808]: I1002 17:04:14.714071 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e714f42f509ea979f837ecd4e70b3cf082c7f64accd50387b1d79e7317f63a30"} err="failed to get container status \"e714f42f509ea979f837ecd4e70b3cf082c7f64accd50387b1d79e7317f63a30\": rpc error: code = NotFound desc = could not find container \"e714f42f509ea979f837ecd4e70b3cf082c7f64accd50387b1d79e7317f63a30\": container with ID starting with e714f42f509ea979f837ecd4e70b3cf082c7f64accd50387b1d79e7317f63a30 not found: ID does not exist" Oct 02 17:04:15 crc kubenswrapper[4808]: I1002 17:04:15.413037 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="843834ff-31a7-4526-95bd-c37bff75cd13" path="/var/lib/kubelet/pods/843834ff-31a7-4526-95bd-c37bff75cd13/volumes" Oct 02 17:04:18 crc kubenswrapper[4808]: I1002 17:04:18.739419 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 17:04:18 crc kubenswrapper[4808]: I1002 17:04:18.739962 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 17:04:19 crc kubenswrapper[4808]: I1002 17:04:19.077171 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-qmfsh_2849754c-63a9-481f-8c8b-f328a0daaa2c/control-plane-machine-set-operator/0.log" Oct 02 17:04:19 crc kubenswrapper[4808]: I1002 17:04:19.271145 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zqnpj_3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e/kube-rbac-proxy/0.log" Oct 02 17:04:19 crc kubenswrapper[4808]: I1002 17:04:19.275544 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zqnpj_3e3da6d6-cdd6-43d2-8910-c0fcd4d19e8e/machine-api-operator/0.log" Oct 02 17:04:31 crc kubenswrapper[4808]: I1002 17:04:31.690804 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-fwdv4_d45bb58b-f832-4f1c-92d0-286ba9ca7ab9/cert-manager-controller/0.log" Oct 02 17:04:31 crc kubenswrapper[4808]: I1002 17:04:31.875267 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-5s7x8_dfc994a0-86f0-4358-a2f3-ff474b5e48d6/cert-manager-cainjector/0.log" Oct 02 17:04:31 crc kubenswrapper[4808]: I1002 17:04:31.891824 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-gp2gc_afc013f6-182d-4d68-bb21-7e933464712b/cert-manager-webhook/0.log" Oct 02 17:04:34 crc kubenswrapper[4808]: I1002 17:04:34.956546 4808 scope.go:117] "RemoveContainer" containerID="28e4a8f6f8eaffa1abe22e2ae1b0be1e3da2c9034034dc4b4e8895d8a6c51d1b" Oct 02 17:04:43 crc kubenswrapper[4808]: I1002 17:04:43.202783 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-96dhm_5684c3ae-4d4a-4c26-b199-1601088fc825/nmstate-console-plugin/0.log" Oct 02 17:04:43 crc kubenswrapper[4808]: I1002 17:04:43.391734 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-gljkg_3ac3fc4a-ef0b-4066-83e6-ebc7d8364629/nmstate-handler/0.log" Oct 02 17:04:43 crc kubenswrapper[4808]: I1002 17:04:43.420484 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-x9xwv_3ee41a88-9b99-4b8c-b985-507fcf86a93e/kube-rbac-proxy/0.log" Oct 02 17:04:43 crc kubenswrapper[4808]: I1002 17:04:43.425264 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-x9xwv_3ee41a88-9b99-4b8c-b985-507fcf86a93e/nmstate-metrics/0.log" Oct 02 17:04:43 crc kubenswrapper[4808]: I1002 17:04:43.598552 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-998ln_b0b5b93b-65b0-4dd5-86a9-2b7151771ff8/nmstate-operator/0.log" Oct 02 17:04:43 crc kubenswrapper[4808]: I1002 17:04:43.665721 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-tr7dz_48ea0327-d95f-42e6-a484-3eb3be93918a/nmstate-webhook/0.log" Oct 02 17:04:48 crc kubenswrapper[4808]: I1002 17:04:48.739495 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 17:04:48 crc kubenswrapper[4808]: I1002 17:04:48.739918 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 17:04:54 crc kubenswrapper[4808]: I1002 17:04:54.893196 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ng8dc"] Oct 02 17:04:54 crc kubenswrapper[4808]: E1002 17:04:54.894530 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="843834ff-31a7-4526-95bd-c37bff75cd13" containerName="registry-server" Oct 02 17:04:54 crc kubenswrapper[4808]: I1002 17:04:54.894556 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="843834ff-31a7-4526-95bd-c37bff75cd13" containerName="registry-server" Oct 02 17:04:54 crc kubenswrapper[4808]: E1002 17:04:54.894589 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="843834ff-31a7-4526-95bd-c37bff75cd13" containerName="extract-utilities" Oct 02 17:04:54 crc kubenswrapper[4808]: I1002 17:04:54.894603 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="843834ff-31a7-4526-95bd-c37bff75cd13" containerName="extract-utilities" Oct 02 17:04:54 crc kubenswrapper[4808]: E1002 17:04:54.894640 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="843834ff-31a7-4526-95bd-c37bff75cd13" containerName="extract-content" Oct 02 17:04:54 crc kubenswrapper[4808]: I1002 17:04:54.894655 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="843834ff-31a7-4526-95bd-c37bff75cd13" containerName="extract-content" Oct 02 17:04:54 crc kubenswrapper[4808]: I1002 17:04:54.894975 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="843834ff-31a7-4526-95bd-c37bff75cd13" containerName="registry-server" Oct 02 17:04:54 crc kubenswrapper[4808]: I1002 17:04:54.896549 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ng8dc" Oct 02 17:04:54 crc kubenswrapper[4808]: I1002 17:04:54.913423 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ng8dc"] Oct 02 17:04:54 crc kubenswrapper[4808]: I1002 17:04:54.990632 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgff8\" (UniqueName: \"kubernetes.io/projected/536dde9e-d4a0-4df5-ae4a-55358853601f-kube-api-access-vgff8\") pod \"community-operators-ng8dc\" (UID: \"536dde9e-d4a0-4df5-ae4a-55358853601f\") " pod="openshift-marketplace/community-operators-ng8dc" Oct 02 17:04:54 crc kubenswrapper[4808]: I1002 17:04:54.990695 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/536dde9e-d4a0-4df5-ae4a-55358853601f-catalog-content\") pod \"community-operators-ng8dc\" (UID: \"536dde9e-d4a0-4df5-ae4a-55358853601f\") " pod="openshift-marketplace/community-operators-ng8dc" Oct 02 17:04:54 crc kubenswrapper[4808]: I1002 17:04:54.991031 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/536dde9e-d4a0-4df5-ae4a-55358853601f-utilities\") pod \"community-operators-ng8dc\" (UID: \"536dde9e-d4a0-4df5-ae4a-55358853601f\") " pod="openshift-marketplace/community-operators-ng8dc" Oct 02 17:04:55 crc kubenswrapper[4808]: I1002 17:04:55.093001 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/536dde9e-d4a0-4df5-ae4a-55358853601f-utilities\") pod \"community-operators-ng8dc\" (UID: \"536dde9e-d4a0-4df5-ae4a-55358853601f\") " pod="openshift-marketplace/community-operators-ng8dc" Oct 02 17:04:55 crc kubenswrapper[4808]: I1002 17:04:55.093175 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgff8\" (UniqueName: \"kubernetes.io/projected/536dde9e-d4a0-4df5-ae4a-55358853601f-kube-api-access-vgff8\") pod \"community-operators-ng8dc\" (UID: \"536dde9e-d4a0-4df5-ae4a-55358853601f\") " pod="openshift-marketplace/community-operators-ng8dc" Oct 02 17:04:55 crc kubenswrapper[4808]: I1002 17:04:55.093207 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/536dde9e-d4a0-4df5-ae4a-55358853601f-catalog-content\") pod \"community-operators-ng8dc\" (UID: \"536dde9e-d4a0-4df5-ae4a-55358853601f\") " pod="openshift-marketplace/community-operators-ng8dc" Oct 02 17:04:55 crc kubenswrapper[4808]: I1002 17:04:55.093635 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/536dde9e-d4a0-4df5-ae4a-55358853601f-utilities\") pod \"community-operators-ng8dc\" (UID: \"536dde9e-d4a0-4df5-ae4a-55358853601f\") " pod="openshift-marketplace/community-operators-ng8dc" Oct 02 17:04:55 crc kubenswrapper[4808]: I1002 17:04:55.093803 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/536dde9e-d4a0-4df5-ae4a-55358853601f-catalog-content\") pod \"community-operators-ng8dc\" (UID: \"536dde9e-d4a0-4df5-ae4a-55358853601f\") " pod="openshift-marketplace/community-operators-ng8dc" Oct 02 17:04:55 crc kubenswrapper[4808]: I1002 17:04:55.115562 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgff8\" (UniqueName: \"kubernetes.io/projected/536dde9e-d4a0-4df5-ae4a-55358853601f-kube-api-access-vgff8\") pod \"community-operators-ng8dc\" (UID: \"536dde9e-d4a0-4df5-ae4a-55358853601f\") " pod="openshift-marketplace/community-operators-ng8dc" Oct 02 17:04:55 crc kubenswrapper[4808]: I1002 17:04:55.234493 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ng8dc" Oct 02 17:04:55 crc kubenswrapper[4808]: I1002 17:04:55.717647 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ng8dc"] Oct 02 17:04:56 crc kubenswrapper[4808]: I1002 17:04:56.042285 4808 generic.go:334] "Generic (PLEG): container finished" podID="536dde9e-d4a0-4df5-ae4a-55358853601f" containerID="51cdd869f6f1974f4ae92ffa1d723f038272a1af071b7542f531d7e00db59985" exitCode=0 Oct 02 17:04:56 crc kubenswrapper[4808]: I1002 17:04:56.043606 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ng8dc" event={"ID":"536dde9e-d4a0-4df5-ae4a-55358853601f","Type":"ContainerDied","Data":"51cdd869f6f1974f4ae92ffa1d723f038272a1af071b7542f531d7e00db59985"} Oct 02 17:04:56 crc kubenswrapper[4808]: I1002 17:04:56.043702 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ng8dc" event={"ID":"536dde9e-d4a0-4df5-ae4a-55358853601f","Type":"ContainerStarted","Data":"abd10efc74f48c4c4c74bbe60a86e67ba8bcd3d326e708754beced76dae8ab2e"} Oct 02 17:04:57 crc kubenswrapper[4808]: I1002 17:04:57.548175 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-57bdp_575f788d-08ad-4138-82a1-fa7fa1543845/kube-rbac-proxy/0.log" Oct 02 17:04:57 crc kubenswrapper[4808]: I1002 17:04:57.702670 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t9q4f_70e2d411-eb7c-46f9-bf6a-06ed29a0bd27/cp-frr-files/0.log" Oct 02 17:04:57 crc kubenswrapper[4808]: I1002 17:04:57.787507 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-57bdp_575f788d-08ad-4138-82a1-fa7fa1543845/controller/0.log" Oct 02 17:04:57 crc kubenswrapper[4808]: I1002 17:04:57.897577 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t9q4f_70e2d411-eb7c-46f9-bf6a-06ed29a0bd27/cp-reloader/0.log" Oct 02 17:04:57 crc kubenswrapper[4808]: I1002 17:04:57.898340 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t9q4f_70e2d411-eb7c-46f9-bf6a-06ed29a0bd27/cp-frr-files/0.log" Oct 02 17:04:57 crc kubenswrapper[4808]: I1002 17:04:57.945715 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t9q4f_70e2d411-eb7c-46f9-bf6a-06ed29a0bd27/cp-metrics/0.log" Oct 02 17:04:57 crc kubenswrapper[4808]: I1002 17:04:57.974060 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t9q4f_70e2d411-eb7c-46f9-bf6a-06ed29a0bd27/cp-reloader/0.log" Oct 02 17:04:58 crc kubenswrapper[4808]: I1002 17:04:58.074385 4808 generic.go:334] "Generic (PLEG): container finished" podID="536dde9e-d4a0-4df5-ae4a-55358853601f" containerID="df115a8f808812a05180f681ba7fdd5d2c73959fcb73922fc70ba825368e79f6" exitCode=0 Oct 02 17:04:58 crc kubenswrapper[4808]: I1002 17:04:58.074428 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ng8dc" event={"ID":"536dde9e-d4a0-4df5-ae4a-55358853601f","Type":"ContainerDied","Data":"df115a8f808812a05180f681ba7fdd5d2c73959fcb73922fc70ba825368e79f6"} Oct 02 17:04:58 crc kubenswrapper[4808]: I1002 17:04:58.176935 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t9q4f_70e2d411-eb7c-46f9-bf6a-06ed29a0bd27/cp-reloader/0.log" Oct 02 17:04:58 crc kubenswrapper[4808]: I1002 17:04:58.178151 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t9q4f_70e2d411-eb7c-46f9-bf6a-06ed29a0bd27/cp-frr-files/0.log" Oct 02 17:04:58 crc kubenswrapper[4808]: I1002 17:04:58.180045 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t9q4f_70e2d411-eb7c-46f9-bf6a-06ed29a0bd27/cp-metrics/0.log" Oct 02 17:04:58 crc kubenswrapper[4808]: I1002 17:04:58.213023 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t9q4f_70e2d411-eb7c-46f9-bf6a-06ed29a0bd27/cp-metrics/0.log" Oct 02 17:04:58 crc kubenswrapper[4808]: I1002 17:04:58.375299 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t9q4f_70e2d411-eb7c-46f9-bf6a-06ed29a0bd27/cp-metrics/0.log" Oct 02 17:04:58 crc kubenswrapper[4808]: I1002 17:04:58.381488 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t9q4f_70e2d411-eb7c-46f9-bf6a-06ed29a0bd27/cp-frr-files/0.log" Oct 02 17:04:58 crc kubenswrapper[4808]: I1002 17:04:58.383487 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t9q4f_70e2d411-eb7c-46f9-bf6a-06ed29a0bd27/cp-reloader/0.log" Oct 02 17:04:58 crc kubenswrapper[4808]: I1002 17:04:58.406001 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t9q4f_70e2d411-eb7c-46f9-bf6a-06ed29a0bd27/controller/0.log" Oct 02 17:04:58 crc kubenswrapper[4808]: I1002 17:04:58.554689 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t9q4f_70e2d411-eb7c-46f9-bf6a-06ed29a0bd27/kube-rbac-proxy/0.log" Oct 02 17:04:58 crc kubenswrapper[4808]: I1002 17:04:58.566815 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t9q4f_70e2d411-eb7c-46f9-bf6a-06ed29a0bd27/frr-metrics/0.log" Oct 02 17:04:58 crc kubenswrapper[4808]: I1002 17:04:58.630433 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t9q4f_70e2d411-eb7c-46f9-bf6a-06ed29a0bd27/kube-rbac-proxy-frr/0.log" Oct 02 17:04:58 crc kubenswrapper[4808]: I1002 17:04:58.763610 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t9q4f_70e2d411-eb7c-46f9-bf6a-06ed29a0bd27/reloader/0.log" Oct 02 17:04:58 crc kubenswrapper[4808]: I1002 17:04:58.835354 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-d5vzc_14914608-ea0e-4a8f-9bb7-88ff2aa55d12/frr-k8s-webhook-server/0.log" Oct 02 17:04:59 crc kubenswrapper[4808]: I1002 17:04:59.061278 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5d89bdf8b7-wfkk6_b8be9f85-599f-4e47-99f5-c4e3d9ef27e9/manager/0.log" Oct 02 17:04:59 crc kubenswrapper[4808]: I1002 17:04:59.084355 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ng8dc" event={"ID":"536dde9e-d4a0-4df5-ae4a-55358853601f","Type":"ContainerStarted","Data":"27e8bc74047b92515408ddda7273a500d3bbcc1020d5934baafd780fd5040bd1"} Oct 02 17:04:59 crc kubenswrapper[4808]: I1002 17:04:59.105555 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ng8dc" podStartSLOduration=2.380601721 podStartE2EDuration="5.105539461s" podCreationTimestamp="2025-10-02 17:04:54 +0000 UTC" firstStartedPulling="2025-10-02 17:04:56.052426925 +0000 UTC m=+1463.377955945" lastFinishedPulling="2025-10-02 17:04:58.777364685 +0000 UTC m=+1466.102893685" observedRunningTime="2025-10-02 17:04:59.100932717 +0000 UTC m=+1466.426461717" watchObservedRunningTime="2025-10-02 17:04:59.105539461 +0000 UTC m=+1466.431068461" Oct 02 17:04:59 crc kubenswrapper[4808]: I1002 17:04:59.236013 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-648b974cdc-8fj4z_03782327-db08-4ac4-a753-2a23156d1329/webhook-server/0.log" Oct 02 17:04:59 crc kubenswrapper[4808]: I1002 17:04:59.330170 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-n9qkx_07aeac15-c264-4b20-ae80-d8d367224849/kube-rbac-proxy/0.log" Oct 02 17:04:59 crc kubenswrapper[4808]: I1002 17:04:59.376837 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t9q4f_70e2d411-eb7c-46f9-bf6a-06ed29a0bd27/frr/0.log" Oct 02 17:04:59 crc kubenswrapper[4808]: I1002 17:04:59.869470 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-n9qkx_07aeac15-c264-4b20-ae80-d8d367224849/speaker/0.log" Oct 02 17:05:05 crc kubenswrapper[4808]: I1002 17:05:05.234854 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ng8dc" Oct 02 17:05:05 crc kubenswrapper[4808]: I1002 17:05:05.235491 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ng8dc" Oct 02 17:05:05 crc kubenswrapper[4808]: I1002 17:05:05.291367 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ng8dc" Oct 02 17:05:06 crc kubenswrapper[4808]: I1002 17:05:06.205560 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ng8dc" Oct 02 17:05:06 crc kubenswrapper[4808]: I1002 17:05:06.260775 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ng8dc"] Oct 02 17:05:08 crc kubenswrapper[4808]: I1002 17:05:08.166255 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ng8dc" podUID="536dde9e-d4a0-4df5-ae4a-55358853601f" containerName="registry-server" containerID="cri-o://27e8bc74047b92515408ddda7273a500d3bbcc1020d5934baafd780fd5040bd1" gracePeriod=2 Oct 02 17:05:10 crc kubenswrapper[4808]: I1002 17:05:10.185218 4808 generic.go:334] "Generic (PLEG): container finished" podID="536dde9e-d4a0-4df5-ae4a-55358853601f" containerID="27e8bc74047b92515408ddda7273a500d3bbcc1020d5934baafd780fd5040bd1" exitCode=0 Oct 02 17:05:10 crc kubenswrapper[4808]: I1002 17:05:10.185275 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ng8dc" event={"ID":"536dde9e-d4a0-4df5-ae4a-55358853601f","Type":"ContainerDied","Data":"27e8bc74047b92515408ddda7273a500d3bbcc1020d5934baafd780fd5040bd1"} Oct 02 17:05:10 crc kubenswrapper[4808]: I1002 17:05:10.442289 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ng8dc" Oct 02 17:05:10 crc kubenswrapper[4808]: I1002 17:05:10.492662 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/536dde9e-d4a0-4df5-ae4a-55358853601f-catalog-content\") pod \"536dde9e-d4a0-4df5-ae4a-55358853601f\" (UID: \"536dde9e-d4a0-4df5-ae4a-55358853601f\") " Oct 02 17:05:10 crc kubenswrapper[4808]: I1002 17:05:10.492868 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/536dde9e-d4a0-4df5-ae4a-55358853601f-utilities\") pod \"536dde9e-d4a0-4df5-ae4a-55358853601f\" (UID: \"536dde9e-d4a0-4df5-ae4a-55358853601f\") " Oct 02 17:05:10 crc kubenswrapper[4808]: I1002 17:05:10.492915 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgff8\" (UniqueName: \"kubernetes.io/projected/536dde9e-d4a0-4df5-ae4a-55358853601f-kube-api-access-vgff8\") pod \"536dde9e-d4a0-4df5-ae4a-55358853601f\" (UID: \"536dde9e-d4a0-4df5-ae4a-55358853601f\") " Oct 02 17:05:10 crc kubenswrapper[4808]: I1002 17:05:10.493866 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/536dde9e-d4a0-4df5-ae4a-55358853601f-utilities" (OuterVolumeSpecName: "utilities") pod "536dde9e-d4a0-4df5-ae4a-55358853601f" (UID: "536dde9e-d4a0-4df5-ae4a-55358853601f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 17:05:10 crc kubenswrapper[4808]: I1002 17:05:10.498767 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/536dde9e-d4a0-4df5-ae4a-55358853601f-kube-api-access-vgff8" (OuterVolumeSpecName: "kube-api-access-vgff8") pod "536dde9e-d4a0-4df5-ae4a-55358853601f" (UID: "536dde9e-d4a0-4df5-ae4a-55358853601f"). InnerVolumeSpecName "kube-api-access-vgff8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:05:10 crc kubenswrapper[4808]: I1002 17:05:10.550121 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/536dde9e-d4a0-4df5-ae4a-55358853601f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "536dde9e-d4a0-4df5-ae4a-55358853601f" (UID: "536dde9e-d4a0-4df5-ae4a-55358853601f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 17:05:10 crc kubenswrapper[4808]: I1002 17:05:10.594965 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/536dde9e-d4a0-4df5-ae4a-55358853601f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 17:05:10 crc kubenswrapper[4808]: I1002 17:05:10.594996 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/536dde9e-d4a0-4df5-ae4a-55358853601f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 17:05:10 crc kubenswrapper[4808]: I1002 17:05:10.595006 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgff8\" (UniqueName: \"kubernetes.io/projected/536dde9e-d4a0-4df5-ae4a-55358853601f-kube-api-access-vgff8\") on node \"crc\" DevicePath \"\"" Oct 02 17:05:11 crc kubenswrapper[4808]: I1002 17:05:11.195918 4808 generic.go:334] "Generic (PLEG): container finished" podID="40e44450-b30c-4df1-b375-ac76b8ae7719" containerID="5bb64046286c402dd492725c73918916c21a02daf5f25fb1268b0a9a82f88d03" exitCode=0 Oct 02 17:05:11 crc kubenswrapper[4808]: I1002 17:05:11.195962 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" event={"ID":"40e44450-b30c-4df1-b375-ac76b8ae7719","Type":"ContainerDied","Data":"5bb64046286c402dd492725c73918916c21a02daf5f25fb1268b0a9a82f88d03"} Oct 02 17:05:11 crc kubenswrapper[4808]: I1002 17:05:11.198858 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ng8dc" event={"ID":"536dde9e-d4a0-4df5-ae4a-55358853601f","Type":"ContainerDied","Data":"abd10efc74f48c4c4c74bbe60a86e67ba8bcd3d326e708754beced76dae8ab2e"} Oct 02 17:05:11 crc kubenswrapper[4808]: I1002 17:05:11.198904 4808 scope.go:117] "RemoveContainer" containerID="27e8bc74047b92515408ddda7273a500d3bbcc1020d5934baafd780fd5040bd1" Oct 02 17:05:11 crc kubenswrapper[4808]: I1002 17:05:11.199026 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ng8dc" Oct 02 17:05:11 crc kubenswrapper[4808]: I1002 17:05:11.225016 4808 scope.go:117] "RemoveContainer" containerID="df115a8f808812a05180f681ba7fdd5d2c73959fcb73922fc70ba825368e79f6" Oct 02 17:05:11 crc kubenswrapper[4808]: I1002 17:05:11.253380 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ng8dc"] Oct 02 17:05:11 crc kubenswrapper[4808]: I1002 17:05:11.267747 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ng8dc"] Oct 02 17:05:11 crc kubenswrapper[4808]: I1002 17:05:11.269381 4808 scope.go:117] "RemoveContainer" containerID="51cdd869f6f1974f4ae92ffa1d723f038272a1af071b7542f531d7e00db59985" Oct 02 17:05:11 crc kubenswrapper[4808]: I1002 17:05:11.406478 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="536dde9e-d4a0-4df5-ae4a-55358853601f" path="/var/lib/kubelet/pods/536dde9e-d4a0-4df5-ae4a-55358853601f/volumes" Oct 02 17:05:11 crc kubenswrapper[4808]: I1002 17:05:11.531038 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl_f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6/util/0.log" Oct 02 17:05:11 crc kubenswrapper[4808]: I1002 17:05:11.692131 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl_f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6/util/0.log" Oct 02 17:05:11 crc kubenswrapper[4808]: I1002 17:05:11.740479 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl_f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6/pull/0.log" Oct 02 17:05:11 crc kubenswrapper[4808]: I1002 17:05:11.743692 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl_f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6/pull/0.log" Oct 02 17:05:11 crc kubenswrapper[4808]: I1002 17:05:11.909911 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl_f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6/extract/0.log" Oct 02 17:05:11 crc kubenswrapper[4808]: I1002 17:05:11.941372 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl_f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6/pull/0.log" Oct 02 17:05:11 crc kubenswrapper[4808]: I1002 17:05:11.988146 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d264jrl_f62ea62e-c2f0-4fa5-8b75-a6b482d6fee6/util/0.log" Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.085486 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nbjkh_6893a3db-ff37-480f-9883-af8db609de88/extract-utilities/0.log" Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.232924 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nbjkh_6893a3db-ff37-480f-9883-af8db609de88/extract-utilities/0.log" Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.243227 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nbjkh_6893a3db-ff37-480f-9883-af8db609de88/extract-content/0.log" Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.248824 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nbjkh_6893a3db-ff37-480f-9883-af8db609de88/extract-content/0.log" Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.452560 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nbjkh_6893a3db-ff37-480f-9883-af8db609de88/extract-content/0.log" Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.487769 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nbjkh_6893a3db-ff37-480f-9883-af8db609de88/extract-utilities/0.log" Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.566039 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.699350 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4q65f_b7ae49e8-10fa-4144-aa62-30d63a77ab0e/extract-utilities/0.log" Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.728715 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e44450-b30c-4df1-b375-ac76b8ae7719-bootstrap-combined-ca-bundle\") pod \"40e44450-b30c-4df1-b375-ac76b8ae7719\" (UID: \"40e44450-b30c-4df1-b375-ac76b8ae7719\") " Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.728772 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/40e44450-b30c-4df1-b375-ac76b8ae7719-inventory\") pod \"40e44450-b30c-4df1-b375-ac76b8ae7719\" (UID: \"40e44450-b30c-4df1-b375-ac76b8ae7719\") " Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.728808 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/40e44450-b30c-4df1-b375-ac76b8ae7719-ssh-key\") pod \"40e44450-b30c-4df1-b375-ac76b8ae7719\" (UID: \"40e44450-b30c-4df1-b375-ac76b8ae7719\") " Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.728882 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txcsd\" (UniqueName: \"kubernetes.io/projected/40e44450-b30c-4df1-b375-ac76b8ae7719-kube-api-access-txcsd\") pod \"40e44450-b30c-4df1-b375-ac76b8ae7719\" (UID: \"40e44450-b30c-4df1-b375-ac76b8ae7719\") " Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.733743 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nbjkh_6893a3db-ff37-480f-9883-af8db609de88/registry-server/0.log" Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.737614 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40e44450-b30c-4df1-b375-ac76b8ae7719-kube-api-access-txcsd" (OuterVolumeSpecName: "kube-api-access-txcsd") pod "40e44450-b30c-4df1-b375-ac76b8ae7719" (UID: "40e44450-b30c-4df1-b375-ac76b8ae7719"). InnerVolumeSpecName "kube-api-access-txcsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.738789 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e44450-b30c-4df1-b375-ac76b8ae7719-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "40e44450-b30c-4df1-b375-ac76b8ae7719" (UID: "40e44450-b30c-4df1-b375-ac76b8ae7719"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.757518 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e44450-b30c-4df1-b375-ac76b8ae7719-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "40e44450-b30c-4df1-b375-ac76b8ae7719" (UID: "40e44450-b30c-4df1-b375-ac76b8ae7719"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.759287 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e44450-b30c-4df1-b375-ac76b8ae7719-inventory" (OuterVolumeSpecName: "inventory") pod "40e44450-b30c-4df1-b375-ac76b8ae7719" (UID: "40e44450-b30c-4df1-b375-ac76b8ae7719"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.830702 4808 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e44450-b30c-4df1-b375-ac76b8ae7719-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.830748 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/40e44450-b30c-4df1-b375-ac76b8ae7719-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.830759 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/40e44450-b30c-4df1-b375-ac76b8ae7719-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 17:05:12 crc kubenswrapper[4808]: I1002 17:05:12.830771 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txcsd\" (UniqueName: \"kubernetes.io/projected/40e44450-b30c-4df1-b375-ac76b8ae7719-kube-api-access-txcsd\") on node \"crc\" DevicePath \"\"" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:12.846566 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4q65f_b7ae49e8-10fa-4144-aa62-30d63a77ab0e/extract-content/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:12.858243 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4q65f_b7ae49e8-10fa-4144-aa62-30d63a77ab0e/extract-utilities/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:12.922627 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4q65f_b7ae49e8-10fa-4144-aa62-30d63a77ab0e/extract-content/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.055370 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4q65f_b7ae49e8-10fa-4144-aa62-30d63a77ab0e/extract-utilities/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.071781 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4q65f_b7ae49e8-10fa-4144-aa62-30d63a77ab0e/extract-content/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.225533 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" event={"ID":"40e44450-b30c-4df1-b375-ac76b8ae7719","Type":"ContainerDied","Data":"b0ce94799e09ec7b1ecce9f6708ecc92a791fc3551641b6309c8966e91a9e547"} Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.225566 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0ce94799e09ec7b1ecce9f6708ecc92a791fc3551641b6309c8966e91a9e547" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.225592 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6ct9" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.292894 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj"] Oct 02 17:05:14 crc kubenswrapper[4808]: E1002 17:05:13.293268 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40e44450-b30c-4df1-b375-ac76b8ae7719" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.293280 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="40e44450-b30c-4df1-b375-ac76b8ae7719" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 17:05:14 crc kubenswrapper[4808]: E1002 17:05:13.293287 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="536dde9e-d4a0-4df5-ae4a-55358853601f" containerName="extract-utilities" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.293293 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="536dde9e-d4a0-4df5-ae4a-55358853601f" containerName="extract-utilities" Oct 02 17:05:14 crc kubenswrapper[4808]: E1002 17:05:13.293317 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="536dde9e-d4a0-4df5-ae4a-55358853601f" containerName="registry-server" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.293324 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="536dde9e-d4a0-4df5-ae4a-55358853601f" containerName="registry-server" Oct 02 17:05:14 crc kubenswrapper[4808]: E1002 17:05:13.293345 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="536dde9e-d4a0-4df5-ae4a-55358853601f" containerName="extract-content" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.293351 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="536dde9e-d4a0-4df5-ae4a-55358853601f" containerName="extract-content" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.293520 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="40e44450-b30c-4df1-b375-ac76b8ae7719" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.293529 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="536dde9e-d4a0-4df5-ae4a-55358853601f" containerName="registry-server" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.294075 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.297358 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.297645 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6w9jl" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.299019 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.301404 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.329094 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj"] Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.368656 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4q65f_b7ae49e8-10fa-4144-aa62-30d63a77ab0e/registry-server/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.443003 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acdbb374-45fe-4396-b387-80f5c2df27e7-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj\" (UID: \"acdbb374-45fe-4396-b387-80f5c2df27e7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.443299 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l52p9\" (UniqueName: \"kubernetes.io/projected/acdbb374-45fe-4396-b387-80f5c2df27e7-kube-api-access-l52p9\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj\" (UID: \"acdbb374-45fe-4396-b387-80f5c2df27e7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.443568 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/acdbb374-45fe-4396-b387-80f5c2df27e7-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj\" (UID: \"acdbb374-45fe-4396-b387-80f5c2df27e7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.544300 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l52p9\" (UniqueName: \"kubernetes.io/projected/acdbb374-45fe-4396-b387-80f5c2df27e7-kube-api-access-l52p9\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj\" (UID: \"acdbb374-45fe-4396-b387-80f5c2df27e7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.544405 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/acdbb374-45fe-4396-b387-80f5c2df27e7-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj\" (UID: \"acdbb374-45fe-4396-b387-80f5c2df27e7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.544449 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acdbb374-45fe-4396-b387-80f5c2df27e7-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj\" (UID: \"acdbb374-45fe-4396-b387-80f5c2df27e7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.550012 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/acdbb374-45fe-4396-b387-80f5c2df27e7-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj\" (UID: \"acdbb374-45fe-4396-b387-80f5c2df27e7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.550074 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acdbb374-45fe-4396-b387-80f5c2df27e7-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj\" (UID: \"acdbb374-45fe-4396-b387-80f5c2df27e7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.570903 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l52p9\" (UniqueName: \"kubernetes.io/projected/acdbb374-45fe-4396-b387-80f5c2df27e7-kube-api-access-l52p9\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj\" (UID: \"acdbb374-45fe-4396-b387-80f5c2df27e7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.610647 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.785501 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5_c19df091-49ac-4606-9cf6-8e9108451371/util/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.908641 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5_c19df091-49ac-4606-9cf6-8e9108451371/util/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.983678 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5_c19df091-49ac-4606-9cf6-8e9108451371/pull/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:13.984019 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5_c19df091-49ac-4606-9cf6-8e9108451371/pull/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:14.180112 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5_c19df091-49ac-4606-9cf6-8e9108451371/util/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:14.207000 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5_c19df091-49ac-4606-9cf6-8e9108451371/extract/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:14.207435 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjcjt5_c19df091-49ac-4606-9cf6-8e9108451371/pull/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:14.378152 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-bhdpb_037881d6-84ee-47ba-a0cf-f6ac68f7ef17/marketplace-operator/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:14.428424 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k8j2q_3f88e6e2-0f03-4f48-8040-b598130fa460/extract-utilities/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:14.664976 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k8j2q_3f88e6e2-0f03-4f48-8040-b598130fa460/extract-content/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:14.665748 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k8j2q_3f88e6e2-0f03-4f48-8040-b598130fa460/extract-content/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:14.700304 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k8j2q_3f88e6e2-0f03-4f48-8040-b598130fa460/extract-utilities/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:14.849900 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj"] Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:14.885530 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k8j2q_3f88e6e2-0f03-4f48-8040-b598130fa460/extract-content/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:14.888469 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k8j2q_3f88e6e2-0f03-4f48-8040-b598130fa460/extract-utilities/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:14.946554 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6gnm_3057619c-ad06-40e4-905b-981c2f7cf6d4/extract-utilities/0.log" Oct 02 17:05:14 crc kubenswrapper[4808]: I1002 17:05:14.985641 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k8j2q_3f88e6e2-0f03-4f48-8040-b598130fa460/registry-server/0.log" Oct 02 17:05:15 crc kubenswrapper[4808]: I1002 17:05:15.115288 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6gnm_3057619c-ad06-40e4-905b-981c2f7cf6d4/extract-utilities/0.log" Oct 02 17:05:15 crc kubenswrapper[4808]: I1002 17:05:15.130804 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6gnm_3057619c-ad06-40e4-905b-981c2f7cf6d4/extract-content/0.log" Oct 02 17:05:15 crc kubenswrapper[4808]: I1002 17:05:15.165690 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6gnm_3057619c-ad06-40e4-905b-981c2f7cf6d4/extract-content/0.log" Oct 02 17:05:15 crc kubenswrapper[4808]: I1002 17:05:15.248226 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj" event={"ID":"acdbb374-45fe-4396-b387-80f5c2df27e7","Type":"ContainerStarted","Data":"3323691167c41dff80b67bc6347da564a0129b538c909fa182e44175c34a6fad"} Oct 02 17:05:15 crc kubenswrapper[4808]: I1002 17:05:15.301933 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6gnm_3057619c-ad06-40e4-905b-981c2f7cf6d4/extract-utilities/0.log" Oct 02 17:05:15 crc kubenswrapper[4808]: I1002 17:05:15.313646 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6gnm_3057619c-ad06-40e4-905b-981c2f7cf6d4/extract-content/0.log" Oct 02 17:05:15 crc kubenswrapper[4808]: I1002 17:05:15.842798 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6gnm_3057619c-ad06-40e4-905b-981c2f7cf6d4/registry-server/0.log" Oct 02 17:05:16 crc kubenswrapper[4808]: I1002 17:05:16.261062 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj" event={"ID":"acdbb374-45fe-4396-b387-80f5c2df27e7","Type":"ContainerStarted","Data":"296a7705df0f693c371caed37e3a648a0bf0bfff6c1f737cfc1632c6f533cc14"} Oct 02 17:05:16 crc kubenswrapper[4808]: I1002 17:05:16.282727 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj" podStartSLOduration=2.299744286 podStartE2EDuration="3.282698772s" podCreationTimestamp="2025-10-02 17:05:13 +0000 UTC" firstStartedPulling="2025-10-02 17:05:14.854637069 +0000 UTC m=+1482.180166069" lastFinishedPulling="2025-10-02 17:05:15.837591535 +0000 UTC m=+1483.163120555" observedRunningTime="2025-10-02 17:05:16.279021633 +0000 UTC m=+1483.604550693" watchObservedRunningTime="2025-10-02 17:05:16.282698772 +0000 UTC m=+1483.608227812" Oct 02 17:05:18 crc kubenswrapper[4808]: I1002 17:05:18.739840 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 17:05:18 crc kubenswrapper[4808]: I1002 17:05:18.740210 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 17:05:18 crc kubenswrapper[4808]: I1002 17:05:18.740314 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 17:05:18 crc kubenswrapper[4808]: I1002 17:05:18.741022 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0fcb0d9bbc76e2f98d002f08b3996525b0e799fbb853c38130d2b118c79a3da9"} pod="openshift-machine-config-operator/machine-config-daemon-7z66r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 17:05:18 crc kubenswrapper[4808]: I1002 17:05:18.741081 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" containerID="cri-o://0fcb0d9bbc76e2f98d002f08b3996525b0e799fbb853c38130d2b118c79a3da9" gracePeriod=600 Oct 02 17:05:19 crc kubenswrapper[4808]: I1002 17:05:19.288336 4808 generic.go:334] "Generic (PLEG): container finished" podID="f83ce425-101d-4489-94a4-5c256eb29328" containerID="0fcb0d9bbc76e2f98d002f08b3996525b0e799fbb853c38130d2b118c79a3da9" exitCode=0 Oct 02 17:05:19 crc kubenswrapper[4808]: I1002 17:05:19.288442 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" event={"ID":"f83ce425-101d-4489-94a4-5c256eb29328","Type":"ContainerDied","Data":"0fcb0d9bbc76e2f98d002f08b3996525b0e799fbb853c38130d2b118c79a3da9"} Oct 02 17:05:19 crc kubenswrapper[4808]: I1002 17:05:19.288890 4808 scope.go:117] "RemoveContainer" containerID="325982695f20979226cf6b1b8bfe2ce4f159293f5eb31e61d627777c496fdb01" Oct 02 17:05:20 crc kubenswrapper[4808]: I1002 17:05:20.311716 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" event={"ID":"f83ce425-101d-4489-94a4-5c256eb29328","Type":"ContainerStarted","Data":"a823e8cfe9d799300a4fab46e28dedc29e9fa6954d8fd2945bf7c266df1f72ba"} Oct 02 17:05:39 crc kubenswrapper[4808]: I1002 17:05:39.471509 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sdqhj"] Oct 02 17:05:39 crc kubenswrapper[4808]: I1002 17:05:39.473630 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sdqhj" Oct 02 17:05:39 crc kubenswrapper[4808]: I1002 17:05:39.482138 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sdqhj"] Oct 02 17:05:39 crc kubenswrapper[4808]: I1002 17:05:39.534713 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd83a58b-33e7-4cdc-9909-9cab949f4374-utilities\") pod \"redhat-marketplace-sdqhj\" (UID: \"cd83a58b-33e7-4cdc-9909-9cab949f4374\") " pod="openshift-marketplace/redhat-marketplace-sdqhj" Oct 02 17:05:39 crc kubenswrapper[4808]: I1002 17:05:39.534810 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfpjd\" (UniqueName: \"kubernetes.io/projected/cd83a58b-33e7-4cdc-9909-9cab949f4374-kube-api-access-lfpjd\") pod \"redhat-marketplace-sdqhj\" (UID: \"cd83a58b-33e7-4cdc-9909-9cab949f4374\") " pod="openshift-marketplace/redhat-marketplace-sdqhj" Oct 02 17:05:39 crc kubenswrapper[4808]: I1002 17:05:39.534839 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd83a58b-33e7-4cdc-9909-9cab949f4374-catalog-content\") pod \"redhat-marketplace-sdqhj\" (UID: \"cd83a58b-33e7-4cdc-9909-9cab949f4374\") " pod="openshift-marketplace/redhat-marketplace-sdqhj" Oct 02 17:05:39 crc kubenswrapper[4808]: I1002 17:05:39.636141 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd83a58b-33e7-4cdc-9909-9cab949f4374-utilities\") pod \"redhat-marketplace-sdqhj\" (UID: \"cd83a58b-33e7-4cdc-9909-9cab949f4374\") " pod="openshift-marketplace/redhat-marketplace-sdqhj" Oct 02 17:05:39 crc kubenswrapper[4808]: I1002 17:05:39.636403 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfpjd\" (UniqueName: \"kubernetes.io/projected/cd83a58b-33e7-4cdc-9909-9cab949f4374-kube-api-access-lfpjd\") pod \"redhat-marketplace-sdqhj\" (UID: \"cd83a58b-33e7-4cdc-9909-9cab949f4374\") " pod="openshift-marketplace/redhat-marketplace-sdqhj" Oct 02 17:05:39 crc kubenswrapper[4808]: I1002 17:05:39.636460 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd83a58b-33e7-4cdc-9909-9cab949f4374-catalog-content\") pod \"redhat-marketplace-sdqhj\" (UID: \"cd83a58b-33e7-4cdc-9909-9cab949f4374\") " pod="openshift-marketplace/redhat-marketplace-sdqhj" Oct 02 17:05:39 crc kubenswrapper[4808]: I1002 17:05:39.636611 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd83a58b-33e7-4cdc-9909-9cab949f4374-utilities\") pod \"redhat-marketplace-sdqhj\" (UID: \"cd83a58b-33e7-4cdc-9909-9cab949f4374\") " pod="openshift-marketplace/redhat-marketplace-sdqhj" Oct 02 17:05:39 crc kubenswrapper[4808]: I1002 17:05:39.636874 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd83a58b-33e7-4cdc-9909-9cab949f4374-catalog-content\") pod \"redhat-marketplace-sdqhj\" (UID: \"cd83a58b-33e7-4cdc-9909-9cab949f4374\") " pod="openshift-marketplace/redhat-marketplace-sdqhj" Oct 02 17:05:39 crc kubenswrapper[4808]: I1002 17:05:39.668765 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfpjd\" (UniqueName: \"kubernetes.io/projected/cd83a58b-33e7-4cdc-9909-9cab949f4374-kube-api-access-lfpjd\") pod \"redhat-marketplace-sdqhj\" (UID: \"cd83a58b-33e7-4cdc-9909-9cab949f4374\") " pod="openshift-marketplace/redhat-marketplace-sdqhj" Oct 02 17:05:39 crc kubenswrapper[4808]: I1002 17:05:39.847416 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sdqhj" Oct 02 17:05:39 crc kubenswrapper[4808]: E1002 17:05:39.950979 4808 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.13:58410->38.102.83.13:40061: read tcp 38.102.83.13:58410->38.102.83.13:40061: read: connection reset by peer Oct 02 17:05:40 crc kubenswrapper[4808]: I1002 17:05:40.370153 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sdqhj"] Oct 02 17:05:40 crc kubenswrapper[4808]: I1002 17:05:40.547304 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sdqhj" event={"ID":"cd83a58b-33e7-4cdc-9909-9cab949f4374","Type":"ContainerStarted","Data":"7df0ffcd98a6954e5a9528d2649b21ef154a8c6634dcf35761c76dde6628a4bc"} Oct 02 17:05:41 crc kubenswrapper[4808]: I1002 17:05:41.558134 4808 generic.go:334] "Generic (PLEG): container finished" podID="cd83a58b-33e7-4cdc-9909-9cab949f4374" containerID="492fe432820631713eb94c5dee268dd3d51316ccac8f895a2183c9e8ea7cedb2" exitCode=0 Oct 02 17:05:41 crc kubenswrapper[4808]: I1002 17:05:41.558189 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sdqhj" event={"ID":"cd83a58b-33e7-4cdc-9909-9cab949f4374","Type":"ContainerDied","Data":"492fe432820631713eb94c5dee268dd3d51316ccac8f895a2183c9e8ea7cedb2"} Oct 02 17:05:42 crc kubenswrapper[4808]: I1002 17:05:42.568468 4808 generic.go:334] "Generic (PLEG): container finished" podID="818e02a3-653a-4b36-a68b-9d4e2df0b71e" containerID="f38c1e1eaa31905554cb4c587be088935750e18481e081f3175807eb6c29368c" exitCode=1 Oct 02 17:05:42 crc kubenswrapper[4808]: I1002 17:05:42.568557 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" event={"ID":"818e02a3-653a-4b36-a68b-9d4e2df0b71e","Type":"ContainerDied","Data":"f38c1e1eaa31905554cb4c587be088935750e18481e081f3175807eb6c29368c"} Oct 02 17:05:42 crc kubenswrapper[4808]: I1002 17:05:42.568764 4808 scope.go:117] "RemoveContainer" containerID="dc9639f10a3923cdab18975fdbea062b709d111e09fdf3caa99a2b5ab5ce0a27" Oct 02 17:05:42 crc kubenswrapper[4808]: I1002 17:05:42.569717 4808 scope.go:117] "RemoveContainer" containerID="f38c1e1eaa31905554cb4c587be088935750e18481e081f3175807eb6c29368c" Oct 02 17:05:42 crc kubenswrapper[4808]: E1002 17:05:42.570142 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 40s restarting failed container=manager pod=keystone-operator-controller-manager-65f4bcbb94-svkdd_openstack-operators(818e02a3-653a-4b36-a68b-9d4e2df0b71e)\"" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" podUID="818e02a3-653a-4b36-a68b-9d4e2df0b71e" Oct 02 17:05:42 crc kubenswrapper[4808]: I1002 17:05:42.571301 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sdqhj" event={"ID":"cd83a58b-33e7-4cdc-9909-9cab949f4374","Type":"ContainerStarted","Data":"617f98c162ed5a82a54c49c36fb946bd44dd3d5dd27f592c31bb3385b38dc9da"} Oct 02 17:05:43 crc kubenswrapper[4808]: I1002 17:05:43.581614 4808 generic.go:334] "Generic (PLEG): container finished" podID="cd83a58b-33e7-4cdc-9909-9cab949f4374" containerID="617f98c162ed5a82a54c49c36fb946bd44dd3d5dd27f592c31bb3385b38dc9da" exitCode=0 Oct 02 17:05:43 crc kubenswrapper[4808]: I1002 17:05:43.581682 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sdqhj" event={"ID":"cd83a58b-33e7-4cdc-9909-9cab949f4374","Type":"ContainerDied","Data":"617f98c162ed5a82a54c49c36fb946bd44dd3d5dd27f592c31bb3385b38dc9da"} Oct 02 17:05:44 crc kubenswrapper[4808]: I1002 17:05:44.590812 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sdqhj" event={"ID":"cd83a58b-33e7-4cdc-9909-9cab949f4374","Type":"ContainerStarted","Data":"9f494c9f87fc07c1afb21dce116d2d4137892d8830a760d0371754331ee99395"} Oct 02 17:05:44 crc kubenswrapper[4808]: I1002 17:05:44.618023 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sdqhj" podStartSLOduration=3.015351445 podStartE2EDuration="5.61800648s" podCreationTimestamp="2025-10-02 17:05:39 +0000 UTC" firstStartedPulling="2025-10-02 17:05:41.560460994 +0000 UTC m=+1508.885990014" lastFinishedPulling="2025-10-02 17:05:44.163116059 +0000 UTC m=+1511.488645049" observedRunningTime="2025-10-02 17:05:44.61538881 +0000 UTC m=+1511.940917810" watchObservedRunningTime="2025-10-02 17:05:44.61800648 +0000 UTC m=+1511.943535480" Oct 02 17:05:49 crc kubenswrapper[4808]: I1002 17:05:49.848087 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sdqhj" Oct 02 17:05:49 crc kubenswrapper[4808]: I1002 17:05:49.848673 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sdqhj" Oct 02 17:05:49 crc kubenswrapper[4808]: I1002 17:05:49.916086 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sdqhj" Oct 02 17:05:50 crc kubenswrapper[4808]: I1002 17:05:50.703257 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 17:05:50 crc kubenswrapper[4808]: I1002 17:05:50.704645 4808 scope.go:117] "RemoveContainer" containerID="f38c1e1eaa31905554cb4c587be088935750e18481e081f3175807eb6c29368c" Oct 02 17:05:50 crc kubenswrapper[4808]: E1002 17:05:50.705114 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 40s restarting failed container=manager pod=keystone-operator-controller-manager-65f4bcbb94-svkdd_openstack-operators(818e02a3-653a-4b36-a68b-9d4e2df0b71e)\"" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" podUID="818e02a3-653a-4b36-a68b-9d4e2df0b71e" Oct 02 17:05:50 crc kubenswrapper[4808]: I1002 17:05:50.706318 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sdqhj" Oct 02 17:05:50 crc kubenswrapper[4808]: I1002 17:05:50.788137 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sdqhj"] Oct 02 17:05:52 crc kubenswrapper[4808]: I1002 17:05:52.660283 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sdqhj" podUID="cd83a58b-33e7-4cdc-9909-9cab949f4374" containerName="registry-server" containerID="cri-o://9f494c9f87fc07c1afb21dce116d2d4137892d8830a760d0371754331ee99395" gracePeriod=2 Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.176362 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sdqhj" Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.282211 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfpjd\" (UniqueName: \"kubernetes.io/projected/cd83a58b-33e7-4cdc-9909-9cab949f4374-kube-api-access-lfpjd\") pod \"cd83a58b-33e7-4cdc-9909-9cab949f4374\" (UID: \"cd83a58b-33e7-4cdc-9909-9cab949f4374\") " Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.282312 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd83a58b-33e7-4cdc-9909-9cab949f4374-utilities\") pod \"cd83a58b-33e7-4cdc-9909-9cab949f4374\" (UID: \"cd83a58b-33e7-4cdc-9909-9cab949f4374\") " Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.282377 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd83a58b-33e7-4cdc-9909-9cab949f4374-catalog-content\") pod \"cd83a58b-33e7-4cdc-9909-9cab949f4374\" (UID: \"cd83a58b-33e7-4cdc-9909-9cab949f4374\") " Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.284286 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd83a58b-33e7-4cdc-9909-9cab949f4374-utilities" (OuterVolumeSpecName: "utilities") pod "cd83a58b-33e7-4cdc-9909-9cab949f4374" (UID: "cd83a58b-33e7-4cdc-9909-9cab949f4374"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.294493 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd83a58b-33e7-4cdc-9909-9cab949f4374-kube-api-access-lfpjd" (OuterVolumeSpecName: "kube-api-access-lfpjd") pod "cd83a58b-33e7-4cdc-9909-9cab949f4374" (UID: "cd83a58b-33e7-4cdc-9909-9cab949f4374"). InnerVolumeSpecName "kube-api-access-lfpjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.299669 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd83a58b-33e7-4cdc-9909-9cab949f4374-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd83a58b-33e7-4cdc-9909-9cab949f4374" (UID: "cd83a58b-33e7-4cdc-9909-9cab949f4374"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.386153 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd83a58b-33e7-4cdc-9909-9cab949f4374-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.386210 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd83a58b-33e7-4cdc-9909-9cab949f4374-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.386262 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfpjd\" (UniqueName: \"kubernetes.io/projected/cd83a58b-33e7-4cdc-9909-9cab949f4374-kube-api-access-lfpjd\") on node \"crc\" DevicePath \"\"" Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.676002 4808 generic.go:334] "Generic (PLEG): container finished" podID="cd83a58b-33e7-4cdc-9909-9cab949f4374" containerID="9f494c9f87fc07c1afb21dce116d2d4137892d8830a760d0371754331ee99395" exitCode=0 Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.676079 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sdqhj" event={"ID":"cd83a58b-33e7-4cdc-9909-9cab949f4374","Type":"ContainerDied","Data":"9f494c9f87fc07c1afb21dce116d2d4137892d8830a760d0371754331ee99395"} Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.676151 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sdqhj" event={"ID":"cd83a58b-33e7-4cdc-9909-9cab949f4374","Type":"ContainerDied","Data":"7df0ffcd98a6954e5a9528d2649b21ef154a8c6634dcf35761c76dde6628a4bc"} Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.676184 4808 scope.go:117] "RemoveContainer" containerID="9f494c9f87fc07c1afb21dce116d2d4137892d8830a760d0371754331ee99395" Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.676105 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sdqhj" Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.722013 4808 scope.go:117] "RemoveContainer" containerID="617f98c162ed5a82a54c49c36fb946bd44dd3d5dd27f592c31bb3385b38dc9da" Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.724693 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sdqhj"] Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.736189 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sdqhj"] Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.752198 4808 scope.go:117] "RemoveContainer" containerID="492fe432820631713eb94c5dee268dd3d51316ccac8f895a2183c9e8ea7cedb2" Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.828772 4808 scope.go:117] "RemoveContainer" containerID="9f494c9f87fc07c1afb21dce116d2d4137892d8830a760d0371754331ee99395" Oct 02 17:05:53 crc kubenswrapper[4808]: E1002 17:05:53.829350 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f494c9f87fc07c1afb21dce116d2d4137892d8830a760d0371754331ee99395\": container with ID starting with 9f494c9f87fc07c1afb21dce116d2d4137892d8830a760d0371754331ee99395 not found: ID does not exist" containerID="9f494c9f87fc07c1afb21dce116d2d4137892d8830a760d0371754331ee99395" Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.829407 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f494c9f87fc07c1afb21dce116d2d4137892d8830a760d0371754331ee99395"} err="failed to get container status \"9f494c9f87fc07c1afb21dce116d2d4137892d8830a760d0371754331ee99395\": rpc error: code = NotFound desc = could not find container \"9f494c9f87fc07c1afb21dce116d2d4137892d8830a760d0371754331ee99395\": container with ID starting with 9f494c9f87fc07c1afb21dce116d2d4137892d8830a760d0371754331ee99395 not found: ID does not exist" Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.829444 4808 scope.go:117] "RemoveContainer" containerID="617f98c162ed5a82a54c49c36fb946bd44dd3d5dd27f592c31bb3385b38dc9da" Oct 02 17:05:53 crc kubenswrapper[4808]: E1002 17:05:53.830567 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"617f98c162ed5a82a54c49c36fb946bd44dd3d5dd27f592c31bb3385b38dc9da\": container with ID starting with 617f98c162ed5a82a54c49c36fb946bd44dd3d5dd27f592c31bb3385b38dc9da not found: ID does not exist" containerID="617f98c162ed5a82a54c49c36fb946bd44dd3d5dd27f592c31bb3385b38dc9da" Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.830629 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"617f98c162ed5a82a54c49c36fb946bd44dd3d5dd27f592c31bb3385b38dc9da"} err="failed to get container status \"617f98c162ed5a82a54c49c36fb946bd44dd3d5dd27f592c31bb3385b38dc9da\": rpc error: code = NotFound desc = could not find container \"617f98c162ed5a82a54c49c36fb946bd44dd3d5dd27f592c31bb3385b38dc9da\": container with ID starting with 617f98c162ed5a82a54c49c36fb946bd44dd3d5dd27f592c31bb3385b38dc9da not found: ID does not exist" Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.830668 4808 scope.go:117] "RemoveContainer" containerID="492fe432820631713eb94c5dee268dd3d51316ccac8f895a2183c9e8ea7cedb2" Oct 02 17:05:53 crc kubenswrapper[4808]: E1002 17:05:53.831096 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"492fe432820631713eb94c5dee268dd3d51316ccac8f895a2183c9e8ea7cedb2\": container with ID starting with 492fe432820631713eb94c5dee268dd3d51316ccac8f895a2183c9e8ea7cedb2 not found: ID does not exist" containerID="492fe432820631713eb94c5dee268dd3d51316ccac8f895a2183c9e8ea7cedb2" Oct 02 17:05:53 crc kubenswrapper[4808]: I1002 17:05:53.831159 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"492fe432820631713eb94c5dee268dd3d51316ccac8f895a2183c9e8ea7cedb2"} err="failed to get container status \"492fe432820631713eb94c5dee268dd3d51316ccac8f895a2183c9e8ea7cedb2\": rpc error: code = NotFound desc = could not find container \"492fe432820631713eb94c5dee268dd3d51316ccac8f895a2183c9e8ea7cedb2\": container with ID starting with 492fe432820631713eb94c5dee268dd3d51316ccac8f895a2183c9e8ea7cedb2 not found: ID does not exist" Oct 02 17:05:55 crc kubenswrapper[4808]: I1002 17:05:55.420477 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd83a58b-33e7-4cdc-9909-9cab949f4374" path="/var/lib/kubelet/pods/cd83a58b-33e7-4cdc-9909-9cab949f4374/volumes" Oct 02 17:06:00 crc kubenswrapper[4808]: I1002 17:06:00.703562 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 17:06:00 crc kubenswrapper[4808]: I1002 17:06:00.705371 4808 scope.go:117] "RemoveContainer" containerID="f38c1e1eaa31905554cb4c587be088935750e18481e081f3175807eb6c29368c" Oct 02 17:06:00 crc kubenswrapper[4808]: E1002 17:06:00.705950 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 40s restarting failed container=manager pod=keystone-operator-controller-manager-65f4bcbb94-svkdd_openstack-operators(818e02a3-653a-4b36-a68b-9d4e2df0b71e)\"" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" podUID="818e02a3-653a-4b36-a68b-9d4e2df0b71e" Oct 02 17:06:14 crc kubenswrapper[4808]: I1002 17:06:14.396207 4808 scope.go:117] "RemoveContainer" containerID="f38c1e1eaa31905554cb4c587be088935750e18481e081f3175807eb6c29368c" Oct 02 17:06:14 crc kubenswrapper[4808]: E1002 17:06:14.396986 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 40s restarting failed container=manager pod=keystone-operator-controller-manager-65f4bcbb94-svkdd_openstack-operators(818e02a3-653a-4b36-a68b-9d4e2df0b71e)\"" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" podUID="818e02a3-653a-4b36-a68b-9d4e2df0b71e" Oct 02 17:06:27 crc kubenswrapper[4808]: I1002 17:06:27.069795 4808 generic.go:334] "Generic (PLEG): container finished" podID="acdbb374-45fe-4396-b387-80f5c2df27e7" containerID="296a7705df0f693c371caed37e3a648a0bf0bfff6c1f737cfc1632c6f533cc14" exitCode=0 Oct 02 17:06:27 crc kubenswrapper[4808]: I1002 17:06:27.069896 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj" event={"ID":"acdbb374-45fe-4396-b387-80f5c2df27e7","Type":"ContainerDied","Data":"296a7705df0f693c371caed37e3a648a0bf0bfff6c1f737cfc1632c6f533cc14"} Oct 02 17:06:28 crc kubenswrapper[4808]: I1002 17:06:28.037468 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-rk9zb"] Oct 02 17:06:28 crc kubenswrapper[4808]: I1002 17:06:28.045992 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-rk9zb"] Oct 02 17:06:28 crc kubenswrapper[4808]: I1002 17:06:28.607409 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj" Oct 02 17:06:28 crc kubenswrapper[4808]: I1002 17:06:28.738819 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/acdbb374-45fe-4396-b387-80f5c2df27e7-ssh-key\") pod \"acdbb374-45fe-4396-b387-80f5c2df27e7\" (UID: \"acdbb374-45fe-4396-b387-80f5c2df27e7\") " Oct 02 17:06:28 crc kubenswrapper[4808]: I1002 17:06:28.738914 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l52p9\" (UniqueName: \"kubernetes.io/projected/acdbb374-45fe-4396-b387-80f5c2df27e7-kube-api-access-l52p9\") pod \"acdbb374-45fe-4396-b387-80f5c2df27e7\" (UID: \"acdbb374-45fe-4396-b387-80f5c2df27e7\") " Oct 02 17:06:28 crc kubenswrapper[4808]: I1002 17:06:28.739090 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acdbb374-45fe-4396-b387-80f5c2df27e7-inventory\") pod \"acdbb374-45fe-4396-b387-80f5c2df27e7\" (UID: \"acdbb374-45fe-4396-b387-80f5c2df27e7\") " Oct 02 17:06:28 crc kubenswrapper[4808]: I1002 17:06:28.748475 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acdbb374-45fe-4396-b387-80f5c2df27e7-kube-api-access-l52p9" (OuterVolumeSpecName: "kube-api-access-l52p9") pod "acdbb374-45fe-4396-b387-80f5c2df27e7" (UID: "acdbb374-45fe-4396-b387-80f5c2df27e7"). InnerVolumeSpecName "kube-api-access-l52p9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:06:28 crc kubenswrapper[4808]: I1002 17:06:28.763508 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acdbb374-45fe-4396-b387-80f5c2df27e7-inventory" (OuterVolumeSpecName: "inventory") pod "acdbb374-45fe-4396-b387-80f5c2df27e7" (UID: "acdbb374-45fe-4396-b387-80f5c2df27e7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:06:28 crc kubenswrapper[4808]: I1002 17:06:28.780888 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acdbb374-45fe-4396-b387-80f5c2df27e7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "acdbb374-45fe-4396-b387-80f5c2df27e7" (UID: "acdbb374-45fe-4396-b387-80f5c2df27e7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:06:28 crc kubenswrapper[4808]: I1002 17:06:28.841139 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acdbb374-45fe-4396-b387-80f5c2df27e7-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 17:06:28 crc kubenswrapper[4808]: I1002 17:06:28.841166 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/acdbb374-45fe-4396-b387-80f5c2df27e7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 17:06:28 crc kubenswrapper[4808]: I1002 17:06:28.841176 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l52p9\" (UniqueName: \"kubernetes.io/projected/acdbb374-45fe-4396-b387-80f5c2df27e7-kube-api-access-l52p9\") on node \"crc\" DevicePath \"\"" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.109280 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj" event={"ID":"acdbb374-45fe-4396-b387-80f5c2df27e7","Type":"ContainerDied","Data":"3323691167c41dff80b67bc6347da564a0129b538c909fa182e44175c34a6fad"} Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.109342 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3323691167c41dff80b67bc6347da564a0129b538c909fa182e44175c34a6fad" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.109506 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kxhwj" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.201891 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx"] Oct 02 17:06:29 crc kubenswrapper[4808]: E1002 17:06:29.202447 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd83a58b-33e7-4cdc-9909-9cab949f4374" containerName="extract-utilities" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.202468 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd83a58b-33e7-4cdc-9909-9cab949f4374" containerName="extract-utilities" Oct 02 17:06:29 crc kubenswrapper[4808]: E1002 17:06:29.202497 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd83a58b-33e7-4cdc-9909-9cab949f4374" containerName="extract-content" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.202506 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd83a58b-33e7-4cdc-9909-9cab949f4374" containerName="extract-content" Oct 02 17:06:29 crc kubenswrapper[4808]: E1002 17:06:29.202528 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd83a58b-33e7-4cdc-9909-9cab949f4374" containerName="registry-server" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.202537 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd83a58b-33e7-4cdc-9909-9cab949f4374" containerName="registry-server" Oct 02 17:06:29 crc kubenswrapper[4808]: E1002 17:06:29.202550 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acdbb374-45fe-4396-b387-80f5c2df27e7" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.202559 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="acdbb374-45fe-4396-b387-80f5c2df27e7" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.202804 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="acdbb374-45fe-4396-b387-80f5c2df27e7" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.202834 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd83a58b-33e7-4cdc-9909-9cab949f4374" containerName="registry-server" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.203949 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.206555 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.207568 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6w9jl" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.207696 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.209434 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.215884 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx"] Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.352428 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e769ee71-7c80-415c-981f-5eeebabd7888-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx\" (UID: \"e769ee71-7c80-415c-981f-5eeebabd7888\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.353139 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjkhk\" (UniqueName: \"kubernetes.io/projected/e769ee71-7c80-415c-981f-5eeebabd7888-kube-api-access-qjkhk\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx\" (UID: \"e769ee71-7c80-415c-981f-5eeebabd7888\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.353203 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e769ee71-7c80-415c-981f-5eeebabd7888-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx\" (UID: \"e769ee71-7c80-415c-981f-5eeebabd7888\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.408735 4808 scope.go:117] "RemoveContainer" containerID="f38c1e1eaa31905554cb4c587be088935750e18481e081f3175807eb6c29368c" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.430950 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16cb8757-fcb8-469a-8971-1c6dffa2e3fd" path="/var/lib/kubelet/pods/16cb8757-fcb8-469a-8971-1c6dffa2e3fd/volumes" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.455681 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e769ee71-7c80-415c-981f-5eeebabd7888-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx\" (UID: \"e769ee71-7c80-415c-981f-5eeebabd7888\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.456020 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e769ee71-7c80-415c-981f-5eeebabd7888-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx\" (UID: \"e769ee71-7c80-415c-981f-5eeebabd7888\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.458663 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjkhk\" (UniqueName: \"kubernetes.io/projected/e769ee71-7c80-415c-981f-5eeebabd7888-kube-api-access-qjkhk\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx\" (UID: \"e769ee71-7c80-415c-981f-5eeebabd7888\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.468366 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e769ee71-7c80-415c-981f-5eeebabd7888-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx\" (UID: \"e769ee71-7c80-415c-981f-5eeebabd7888\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.476493 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjkhk\" (UniqueName: \"kubernetes.io/projected/e769ee71-7c80-415c-981f-5eeebabd7888-kube-api-access-qjkhk\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx\" (UID: \"e769ee71-7c80-415c-981f-5eeebabd7888\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.476966 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e769ee71-7c80-415c-981f-5eeebabd7888-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx\" (UID: \"e769ee71-7c80-415c-981f-5eeebabd7888\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx" Oct 02 17:06:29 crc kubenswrapper[4808]: I1002 17:06:29.542965 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx" Oct 02 17:06:30 crc kubenswrapper[4808]: I1002 17:06:30.105110 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx"] Oct 02 17:06:30 crc kubenswrapper[4808]: W1002 17:06:30.115179 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode769ee71_7c80_415c_981f_5eeebabd7888.slice/crio-c55c4a928aa0d4862edc72fb46fde7bfa599f431066f347098b7cea8443c9429 WatchSource:0}: Error finding container c55c4a928aa0d4862edc72fb46fde7bfa599f431066f347098b7cea8443c9429: Status 404 returned error can't find the container with id c55c4a928aa0d4862edc72fb46fde7bfa599f431066f347098b7cea8443c9429 Oct 02 17:06:30 crc kubenswrapper[4808]: I1002 17:06:30.118843 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 17:06:30 crc kubenswrapper[4808]: I1002 17:06:30.129641 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" event={"ID":"818e02a3-653a-4b36-a68b-9d4e2df0b71e","Type":"ContainerStarted","Data":"a02553b70014af81a0fd9e883ea96771d4b4a6fd6b02b442f8fd5258fdd307ed"} Oct 02 17:06:30 crc kubenswrapper[4808]: I1002 17:06:30.131162 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 17:06:31 crc kubenswrapper[4808]: I1002 17:06:31.167625 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx" event={"ID":"e769ee71-7c80-415c-981f-5eeebabd7888","Type":"ContainerStarted","Data":"22ef185611ebe36b65bf7855125c8dd15f7ee0434f70cf9a7a7a372fe3d6b836"} Oct 02 17:06:31 crc kubenswrapper[4808]: I1002 17:06:31.172208 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx" event={"ID":"e769ee71-7c80-415c-981f-5eeebabd7888","Type":"ContainerStarted","Data":"c55c4a928aa0d4862edc72fb46fde7bfa599f431066f347098b7cea8443c9429"} Oct 02 17:06:31 crc kubenswrapper[4808]: I1002 17:06:31.202090 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx" podStartSLOduration=1.704948208 podStartE2EDuration="2.202015076s" podCreationTimestamp="2025-10-02 17:06:29 +0000 UTC" firstStartedPulling="2025-10-02 17:06:30.118493465 +0000 UTC m=+1557.444022495" lastFinishedPulling="2025-10-02 17:06:30.615560343 +0000 UTC m=+1557.941089363" observedRunningTime="2025-10-02 17:06:31.185332382 +0000 UTC m=+1558.510861422" watchObservedRunningTime="2025-10-02 17:06:31.202015076 +0000 UTC m=+1558.527544116" Oct 02 17:06:31 crc kubenswrapper[4808]: I1002 17:06:31.979525 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pv5gf"] Oct 02 17:06:31 crc kubenswrapper[4808]: I1002 17:06:31.983127 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pv5gf" Oct 02 17:06:31 crc kubenswrapper[4808]: I1002 17:06:31.987772 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pv5gf"] Oct 02 17:06:32 crc kubenswrapper[4808]: I1002 17:06:32.042700 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-twjnb"] Oct 02 17:06:32 crc kubenswrapper[4808]: I1002 17:06:32.054545 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-5fthx"] Oct 02 17:06:32 crc kubenswrapper[4808]: I1002 17:06:32.062849 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-twjnb"] Oct 02 17:06:32 crc kubenswrapper[4808]: I1002 17:06:32.071516 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-5fthx"] Oct 02 17:06:32 crc kubenswrapper[4808]: I1002 17:06:32.123819 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc0a2ce9-29ac-476e-8028-a7167f9c5ef0-catalog-content\") pod \"certified-operators-pv5gf\" (UID: \"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0\") " pod="openshift-marketplace/certified-operators-pv5gf" Oct 02 17:06:32 crc kubenswrapper[4808]: I1002 17:06:32.123949 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc0a2ce9-29ac-476e-8028-a7167f9c5ef0-utilities\") pod \"certified-operators-pv5gf\" (UID: \"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0\") " pod="openshift-marketplace/certified-operators-pv5gf" Oct 02 17:06:32 crc kubenswrapper[4808]: I1002 17:06:32.124023 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggw5j\" (UniqueName: \"kubernetes.io/projected/dc0a2ce9-29ac-476e-8028-a7167f9c5ef0-kube-api-access-ggw5j\") pod \"certified-operators-pv5gf\" (UID: \"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0\") " pod="openshift-marketplace/certified-operators-pv5gf" Oct 02 17:06:32 crc kubenswrapper[4808]: I1002 17:06:32.226356 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc0a2ce9-29ac-476e-8028-a7167f9c5ef0-catalog-content\") pod \"certified-operators-pv5gf\" (UID: \"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0\") " pod="openshift-marketplace/certified-operators-pv5gf" Oct 02 17:06:32 crc kubenswrapper[4808]: I1002 17:06:32.226789 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc0a2ce9-29ac-476e-8028-a7167f9c5ef0-utilities\") pod \"certified-operators-pv5gf\" (UID: \"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0\") " pod="openshift-marketplace/certified-operators-pv5gf" Oct 02 17:06:32 crc kubenswrapper[4808]: I1002 17:06:32.226849 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggw5j\" (UniqueName: \"kubernetes.io/projected/dc0a2ce9-29ac-476e-8028-a7167f9c5ef0-kube-api-access-ggw5j\") pod \"certified-operators-pv5gf\" (UID: \"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0\") " pod="openshift-marketplace/certified-operators-pv5gf" Oct 02 17:06:32 crc kubenswrapper[4808]: I1002 17:06:32.227749 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc0a2ce9-29ac-476e-8028-a7167f9c5ef0-catalog-content\") pod \"certified-operators-pv5gf\" (UID: \"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0\") " pod="openshift-marketplace/certified-operators-pv5gf" Oct 02 17:06:32 crc kubenswrapper[4808]: I1002 17:06:32.228117 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc0a2ce9-29ac-476e-8028-a7167f9c5ef0-utilities\") pod \"certified-operators-pv5gf\" (UID: \"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0\") " pod="openshift-marketplace/certified-operators-pv5gf" Oct 02 17:06:32 crc kubenswrapper[4808]: I1002 17:06:32.246814 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggw5j\" (UniqueName: \"kubernetes.io/projected/dc0a2ce9-29ac-476e-8028-a7167f9c5ef0-kube-api-access-ggw5j\") pod \"certified-operators-pv5gf\" (UID: \"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0\") " pod="openshift-marketplace/certified-operators-pv5gf" Oct 02 17:06:32 crc kubenswrapper[4808]: I1002 17:06:32.309523 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pv5gf" Oct 02 17:06:32 crc kubenswrapper[4808]: I1002 17:06:32.812575 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pv5gf"] Oct 02 17:06:33 crc kubenswrapper[4808]: I1002 17:06:33.188637 4808 generic.go:334] "Generic (PLEG): container finished" podID="dc0a2ce9-29ac-476e-8028-a7167f9c5ef0" containerID="7dd6ccc3300b06071d3659aa79c743b63cb130b7b050cb0e88da25c34920e581" exitCode=0 Oct 02 17:06:33 crc kubenswrapper[4808]: I1002 17:06:33.188853 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pv5gf" event={"ID":"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0","Type":"ContainerDied","Data":"7dd6ccc3300b06071d3659aa79c743b63cb130b7b050cb0e88da25c34920e581"} Oct 02 17:06:33 crc kubenswrapper[4808]: I1002 17:06:33.189014 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pv5gf" event={"ID":"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0","Type":"ContainerStarted","Data":"551dbe4dc96883057ed02c7d231286f36cf688a2752236df558217774e113364"} Oct 02 17:06:33 crc kubenswrapper[4808]: I1002 17:06:33.423964 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c" path="/var/lib/kubelet/pods/06f1c7d3-8b90-4c6f-8d1b-c568cfc9f92c/volumes" Oct 02 17:06:33 crc kubenswrapper[4808]: I1002 17:06:33.425279 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b6c4102-3ccd-4119-bbad-697ad2589e40" path="/var/lib/kubelet/pods/0b6c4102-3ccd-4119-bbad-697ad2589e40/volumes" Oct 02 17:06:35 crc kubenswrapper[4808]: I1002 17:06:35.078643 4808 scope.go:117] "RemoveContainer" containerID="1a8aed86891bb916834343bdd5ae0633a486bc125740a2ec7873690cdf85f01a" Oct 02 17:06:35 crc kubenswrapper[4808]: I1002 17:06:35.102941 4808 scope.go:117] "RemoveContainer" containerID="7540bac8b871f290c1c2c2b8ce6c7ebbeace8f4dddb12fe60a0a4225b17e3b29" Oct 02 17:06:35 crc kubenswrapper[4808]: I1002 17:06:35.140841 4808 scope.go:117] "RemoveContainer" containerID="60dbecb997aa51947969f39a54cbe1e672c3a01fe39c6255ff7e2d99445e9de4" Oct 02 17:06:35 crc kubenswrapper[4808]: I1002 17:06:35.222033 4808 generic.go:334] "Generic (PLEG): container finished" podID="dc0a2ce9-29ac-476e-8028-a7167f9c5ef0" containerID="1d7475ff0349f084be1d3c933c521d43f909834be4dc5d8bf7115e22cf721ba8" exitCode=0 Oct 02 17:06:35 crc kubenswrapper[4808]: I1002 17:06:35.222078 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pv5gf" event={"ID":"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0","Type":"ContainerDied","Data":"1d7475ff0349f084be1d3c933c521d43f909834be4dc5d8bf7115e22cf721ba8"} Oct 02 17:06:36 crc kubenswrapper[4808]: I1002 17:06:36.233727 4808 generic.go:334] "Generic (PLEG): container finished" podID="e769ee71-7c80-415c-981f-5eeebabd7888" containerID="22ef185611ebe36b65bf7855125c8dd15f7ee0434f70cf9a7a7a372fe3d6b836" exitCode=0 Oct 02 17:06:36 crc kubenswrapper[4808]: I1002 17:06:36.233811 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx" event={"ID":"e769ee71-7c80-415c-981f-5eeebabd7888","Type":"ContainerDied","Data":"22ef185611ebe36b65bf7855125c8dd15f7ee0434f70cf9a7a7a372fe3d6b836"} Oct 02 17:06:36 crc kubenswrapper[4808]: I1002 17:06:36.237992 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pv5gf" event={"ID":"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0","Type":"ContainerStarted","Data":"fc8f210ff626b9fca9d5ba99f21ca346f989ab176b3f994dfe009c04f14e82ef"} Oct 02 17:06:36 crc kubenswrapper[4808]: I1002 17:06:36.282218 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pv5gf" podStartSLOduration=2.728566437 podStartE2EDuration="5.282202742s" podCreationTimestamp="2025-10-02 17:06:31 +0000 UTC" firstStartedPulling="2025-10-02 17:06:33.19194475 +0000 UTC m=+1560.517473740" lastFinishedPulling="2025-10-02 17:06:35.745581045 +0000 UTC m=+1563.071110045" observedRunningTime="2025-10-02 17:06:36.275767215 +0000 UTC m=+1563.601296215" watchObservedRunningTime="2025-10-02 17:06:36.282202742 +0000 UTC m=+1563.607731742" Oct 02 17:06:37 crc kubenswrapper[4808]: I1002 17:06:37.684667 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx" Oct 02 17:06:37 crc kubenswrapper[4808]: I1002 17:06:37.848401 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e769ee71-7c80-415c-981f-5eeebabd7888-ssh-key\") pod \"e769ee71-7c80-415c-981f-5eeebabd7888\" (UID: \"e769ee71-7c80-415c-981f-5eeebabd7888\") " Oct 02 17:06:37 crc kubenswrapper[4808]: I1002 17:06:37.848539 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e769ee71-7c80-415c-981f-5eeebabd7888-inventory\") pod \"e769ee71-7c80-415c-981f-5eeebabd7888\" (UID: \"e769ee71-7c80-415c-981f-5eeebabd7888\") " Oct 02 17:06:37 crc kubenswrapper[4808]: I1002 17:06:37.848589 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjkhk\" (UniqueName: \"kubernetes.io/projected/e769ee71-7c80-415c-981f-5eeebabd7888-kube-api-access-qjkhk\") pod \"e769ee71-7c80-415c-981f-5eeebabd7888\" (UID: \"e769ee71-7c80-415c-981f-5eeebabd7888\") " Oct 02 17:06:37 crc kubenswrapper[4808]: I1002 17:06:37.866552 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e769ee71-7c80-415c-981f-5eeebabd7888-kube-api-access-qjkhk" (OuterVolumeSpecName: "kube-api-access-qjkhk") pod "e769ee71-7c80-415c-981f-5eeebabd7888" (UID: "e769ee71-7c80-415c-981f-5eeebabd7888"). InnerVolumeSpecName "kube-api-access-qjkhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:06:37 crc kubenswrapper[4808]: I1002 17:06:37.891472 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e769ee71-7c80-415c-981f-5eeebabd7888-inventory" (OuterVolumeSpecName: "inventory") pod "e769ee71-7c80-415c-981f-5eeebabd7888" (UID: "e769ee71-7c80-415c-981f-5eeebabd7888"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:06:37 crc kubenswrapper[4808]: I1002 17:06:37.896151 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e769ee71-7c80-415c-981f-5eeebabd7888-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e769ee71-7c80-415c-981f-5eeebabd7888" (UID: "e769ee71-7c80-415c-981f-5eeebabd7888"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:06:37 crc kubenswrapper[4808]: I1002 17:06:37.950846 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e769ee71-7c80-415c-981f-5eeebabd7888-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 17:06:37 crc kubenswrapper[4808]: I1002 17:06:37.951084 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjkhk\" (UniqueName: \"kubernetes.io/projected/e769ee71-7c80-415c-981f-5eeebabd7888-kube-api-access-qjkhk\") on node \"crc\" DevicePath \"\"" Oct 02 17:06:37 crc kubenswrapper[4808]: I1002 17:06:37.951195 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e769ee71-7c80-415c-981f-5eeebabd7888-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.038961 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-b4e3-account-create-qsp8f"] Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.080985 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-b4e3-account-create-qsp8f"] Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.256415 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx" event={"ID":"e769ee71-7c80-415c-981f-5eeebabd7888","Type":"ContainerDied","Data":"c55c4a928aa0d4862edc72fb46fde7bfa599f431066f347098b7cea8443c9429"} Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.256458 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c55c4a928aa0d4862edc72fb46fde7bfa599f431066f347098b7cea8443c9429" Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.256525 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cpfdx" Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.319384 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq"] Oct 02 17:06:38 crc kubenswrapper[4808]: E1002 17:06:38.319904 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e769ee71-7c80-415c-981f-5eeebabd7888" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.319971 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e769ee71-7c80-415c-981f-5eeebabd7888" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.320204 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e769ee71-7c80-415c-981f-5eeebabd7888" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.320855 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq" Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.324472 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.325751 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.325974 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6w9jl" Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.334937 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq"] Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.337590 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.464762 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a18970b-8022-4e05-8068-77b866f91b82-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-scfqq\" (UID: \"9a18970b-8022-4e05-8068-77b866f91b82\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq" Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.465326 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a18970b-8022-4e05-8068-77b866f91b82-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-scfqq\" (UID: \"9a18970b-8022-4e05-8068-77b866f91b82\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq" Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.465548 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgnb8\" (UniqueName: \"kubernetes.io/projected/9a18970b-8022-4e05-8068-77b866f91b82-kube-api-access-sgnb8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-scfqq\" (UID: \"9a18970b-8022-4e05-8068-77b866f91b82\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq" Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.568002 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a18970b-8022-4e05-8068-77b866f91b82-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-scfqq\" (UID: \"9a18970b-8022-4e05-8068-77b866f91b82\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq" Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.568523 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgnb8\" (UniqueName: \"kubernetes.io/projected/9a18970b-8022-4e05-8068-77b866f91b82-kube-api-access-sgnb8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-scfqq\" (UID: \"9a18970b-8022-4e05-8068-77b866f91b82\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq" Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.568732 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a18970b-8022-4e05-8068-77b866f91b82-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-scfqq\" (UID: \"9a18970b-8022-4e05-8068-77b866f91b82\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq" Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.581802 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a18970b-8022-4e05-8068-77b866f91b82-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-scfqq\" (UID: \"9a18970b-8022-4e05-8068-77b866f91b82\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq" Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.581901 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a18970b-8022-4e05-8068-77b866f91b82-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-scfqq\" (UID: \"9a18970b-8022-4e05-8068-77b866f91b82\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq" Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.586853 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgnb8\" (UniqueName: \"kubernetes.io/projected/9a18970b-8022-4e05-8068-77b866f91b82-kube-api-access-sgnb8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-scfqq\" (UID: \"9a18970b-8022-4e05-8068-77b866f91b82\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq" Oct 02 17:06:38 crc kubenswrapper[4808]: I1002 17:06:38.639572 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq" Oct 02 17:06:39 crc kubenswrapper[4808]: I1002 17:06:39.218358 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq"] Oct 02 17:06:39 crc kubenswrapper[4808]: W1002 17:06:39.226532 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a18970b_8022_4e05_8068_77b866f91b82.slice/crio-8716bcfd614419032db7682e8510436135af131f8a014175f7b76dceef4bf3ce WatchSource:0}: Error finding container 8716bcfd614419032db7682e8510436135af131f8a014175f7b76dceef4bf3ce: Status 404 returned error can't find the container with id 8716bcfd614419032db7682e8510436135af131f8a014175f7b76dceef4bf3ce Oct 02 17:06:39 crc kubenswrapper[4808]: I1002 17:06:39.266615 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq" event={"ID":"9a18970b-8022-4e05-8068-77b866f91b82","Type":"ContainerStarted","Data":"8716bcfd614419032db7682e8510436135af131f8a014175f7b76dceef4bf3ce"} Oct 02 17:06:39 crc kubenswrapper[4808]: I1002 17:06:39.408799 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="387ae2f5-c39e-44a8-b471-64ab91a40d01" path="/var/lib/kubelet/pods/387ae2f5-c39e-44a8-b471-64ab91a40d01/volumes" Oct 02 17:06:40 crc kubenswrapper[4808]: I1002 17:06:40.706299 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 17:06:41 crc kubenswrapper[4808]: I1002 17:06:41.293069 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq" event={"ID":"9a18970b-8022-4e05-8068-77b866f91b82","Type":"ContainerStarted","Data":"7e0ceabd97171ee9db645825cae3cab7ab3300b18eea10a00e19260d4972e829"} Oct 02 17:06:41 crc kubenswrapper[4808]: I1002 17:06:41.329166 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq" podStartSLOduration=2.5507928 podStartE2EDuration="3.329148604s" podCreationTimestamp="2025-10-02 17:06:38 +0000 UTC" firstStartedPulling="2025-10-02 17:06:39.22966375 +0000 UTC m=+1566.555192750" lastFinishedPulling="2025-10-02 17:06:40.008019514 +0000 UTC m=+1567.333548554" observedRunningTime="2025-10-02 17:06:41.316548477 +0000 UTC m=+1568.642077497" watchObservedRunningTime="2025-10-02 17:06:41.329148604 +0000 UTC m=+1568.654677624" Oct 02 17:06:42 crc kubenswrapper[4808]: I1002 17:06:42.045465 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-828c-account-create-6vmr9"] Oct 02 17:06:42 crc kubenswrapper[4808]: I1002 17:06:42.058037 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-3d12-account-create-wqs6d"] Oct 02 17:06:42 crc kubenswrapper[4808]: I1002 17:06:42.065146 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-828c-account-create-6vmr9"] Oct 02 17:06:42 crc kubenswrapper[4808]: I1002 17:06:42.093736 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-3d12-account-create-wqs6d"] Oct 02 17:06:42 crc kubenswrapper[4808]: I1002 17:06:42.311160 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pv5gf" Oct 02 17:06:42 crc kubenswrapper[4808]: I1002 17:06:42.311206 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pv5gf" Oct 02 17:06:42 crc kubenswrapper[4808]: I1002 17:06:42.392892 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pv5gf" Oct 02 17:06:43 crc kubenswrapper[4808]: I1002 17:06:43.359672 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pv5gf" Oct 02 17:06:43 crc kubenswrapper[4808]: I1002 17:06:43.414114 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf2f1836-9826-4d1c-89ea-eb979de32e62" path="/var/lib/kubelet/pods/bf2f1836-9826-4d1c-89ea-eb979de32e62/volumes" Oct 02 17:06:43 crc kubenswrapper[4808]: I1002 17:06:43.414714 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f446b584-b5e2-48fa-9f7d-3b6ccdc92944" path="/var/lib/kubelet/pods/f446b584-b5e2-48fa-9f7d-3b6ccdc92944/volumes" Oct 02 17:06:43 crc kubenswrapper[4808]: I1002 17:06:43.415293 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pv5gf"] Oct 02 17:06:45 crc kubenswrapper[4808]: I1002 17:06:45.327651 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pv5gf" podUID="dc0a2ce9-29ac-476e-8028-a7167f9c5ef0" containerName="registry-server" containerID="cri-o://fc8f210ff626b9fca9d5ba99f21ca346f989ab176b3f994dfe009c04f14e82ef" gracePeriod=2 Oct 02 17:06:45 crc kubenswrapper[4808]: I1002 17:06:45.754926 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pv5gf" Oct 02 17:06:45 crc kubenswrapper[4808]: I1002 17:06:45.914493 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc0a2ce9-29ac-476e-8028-a7167f9c5ef0-utilities\") pod \"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0\" (UID: \"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0\") " Oct 02 17:06:45 crc kubenswrapper[4808]: I1002 17:06:45.914585 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggw5j\" (UniqueName: \"kubernetes.io/projected/dc0a2ce9-29ac-476e-8028-a7167f9c5ef0-kube-api-access-ggw5j\") pod \"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0\" (UID: \"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0\") " Oct 02 17:06:45 crc kubenswrapper[4808]: I1002 17:06:45.914701 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc0a2ce9-29ac-476e-8028-a7167f9c5ef0-catalog-content\") pod \"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0\" (UID: \"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0\") " Oct 02 17:06:45 crc kubenswrapper[4808]: I1002 17:06:45.915524 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc0a2ce9-29ac-476e-8028-a7167f9c5ef0-utilities" (OuterVolumeSpecName: "utilities") pod "dc0a2ce9-29ac-476e-8028-a7167f9c5ef0" (UID: "dc0a2ce9-29ac-476e-8028-a7167f9c5ef0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 17:06:45 crc kubenswrapper[4808]: I1002 17:06:45.921147 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc0a2ce9-29ac-476e-8028-a7167f9c5ef0-kube-api-access-ggw5j" (OuterVolumeSpecName: "kube-api-access-ggw5j") pod "dc0a2ce9-29ac-476e-8028-a7167f9c5ef0" (UID: "dc0a2ce9-29ac-476e-8028-a7167f9c5ef0"). InnerVolumeSpecName "kube-api-access-ggw5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:06:45 crc kubenswrapper[4808]: I1002 17:06:45.963597 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc0a2ce9-29ac-476e-8028-a7167f9c5ef0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc0a2ce9-29ac-476e-8028-a7167f9c5ef0" (UID: "dc0a2ce9-29ac-476e-8028-a7167f9c5ef0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 17:06:46 crc kubenswrapper[4808]: I1002 17:06:46.017051 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggw5j\" (UniqueName: \"kubernetes.io/projected/dc0a2ce9-29ac-476e-8028-a7167f9c5ef0-kube-api-access-ggw5j\") on node \"crc\" DevicePath \"\"" Oct 02 17:06:46 crc kubenswrapper[4808]: I1002 17:06:46.017416 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc0a2ce9-29ac-476e-8028-a7167f9c5ef0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 17:06:46 crc kubenswrapper[4808]: I1002 17:06:46.017634 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc0a2ce9-29ac-476e-8028-a7167f9c5ef0-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 17:06:46 crc kubenswrapper[4808]: I1002 17:06:46.344315 4808 generic.go:334] "Generic (PLEG): container finished" podID="dc0a2ce9-29ac-476e-8028-a7167f9c5ef0" containerID="fc8f210ff626b9fca9d5ba99f21ca346f989ab176b3f994dfe009c04f14e82ef" exitCode=0 Oct 02 17:06:46 crc kubenswrapper[4808]: I1002 17:06:46.344420 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pv5gf" event={"ID":"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0","Type":"ContainerDied","Data":"fc8f210ff626b9fca9d5ba99f21ca346f989ab176b3f994dfe009c04f14e82ef"} Oct 02 17:06:46 crc kubenswrapper[4808]: I1002 17:06:46.344846 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pv5gf" event={"ID":"dc0a2ce9-29ac-476e-8028-a7167f9c5ef0","Type":"ContainerDied","Data":"551dbe4dc96883057ed02c7d231286f36cf688a2752236df558217774e113364"} Oct 02 17:06:46 crc kubenswrapper[4808]: I1002 17:06:46.344891 4808 scope.go:117] "RemoveContainer" containerID="fc8f210ff626b9fca9d5ba99f21ca346f989ab176b3f994dfe009c04f14e82ef" Oct 02 17:06:46 crc kubenswrapper[4808]: I1002 17:06:46.344449 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pv5gf" Oct 02 17:06:46 crc kubenswrapper[4808]: I1002 17:06:46.398074 4808 scope.go:117] "RemoveContainer" containerID="1d7475ff0349f084be1d3c933c521d43f909834be4dc5d8bf7115e22cf721ba8" Oct 02 17:06:46 crc kubenswrapper[4808]: I1002 17:06:46.408194 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pv5gf"] Oct 02 17:06:46 crc kubenswrapper[4808]: I1002 17:06:46.419054 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pv5gf"] Oct 02 17:06:46 crc kubenswrapper[4808]: I1002 17:06:46.435415 4808 scope.go:117] "RemoveContainer" containerID="7dd6ccc3300b06071d3659aa79c743b63cb130b7b050cb0e88da25c34920e581" Oct 02 17:06:46 crc kubenswrapper[4808]: I1002 17:06:46.486156 4808 scope.go:117] "RemoveContainer" containerID="fc8f210ff626b9fca9d5ba99f21ca346f989ab176b3f994dfe009c04f14e82ef" Oct 02 17:06:46 crc kubenswrapper[4808]: E1002 17:06:46.486629 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc8f210ff626b9fca9d5ba99f21ca346f989ab176b3f994dfe009c04f14e82ef\": container with ID starting with fc8f210ff626b9fca9d5ba99f21ca346f989ab176b3f994dfe009c04f14e82ef not found: ID does not exist" containerID="fc8f210ff626b9fca9d5ba99f21ca346f989ab176b3f994dfe009c04f14e82ef" Oct 02 17:06:46 crc kubenswrapper[4808]: I1002 17:06:46.486678 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc8f210ff626b9fca9d5ba99f21ca346f989ab176b3f994dfe009c04f14e82ef"} err="failed to get container status \"fc8f210ff626b9fca9d5ba99f21ca346f989ab176b3f994dfe009c04f14e82ef\": rpc error: code = NotFound desc = could not find container \"fc8f210ff626b9fca9d5ba99f21ca346f989ab176b3f994dfe009c04f14e82ef\": container with ID starting with fc8f210ff626b9fca9d5ba99f21ca346f989ab176b3f994dfe009c04f14e82ef not found: ID does not exist" Oct 02 17:06:46 crc kubenswrapper[4808]: I1002 17:06:46.486710 4808 scope.go:117] "RemoveContainer" containerID="1d7475ff0349f084be1d3c933c521d43f909834be4dc5d8bf7115e22cf721ba8" Oct 02 17:06:46 crc kubenswrapper[4808]: E1002 17:06:46.487141 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d7475ff0349f084be1d3c933c521d43f909834be4dc5d8bf7115e22cf721ba8\": container with ID starting with 1d7475ff0349f084be1d3c933c521d43f909834be4dc5d8bf7115e22cf721ba8 not found: ID does not exist" containerID="1d7475ff0349f084be1d3c933c521d43f909834be4dc5d8bf7115e22cf721ba8" Oct 02 17:06:46 crc kubenswrapper[4808]: I1002 17:06:46.487170 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d7475ff0349f084be1d3c933c521d43f909834be4dc5d8bf7115e22cf721ba8"} err="failed to get container status \"1d7475ff0349f084be1d3c933c521d43f909834be4dc5d8bf7115e22cf721ba8\": rpc error: code = NotFound desc = could not find container \"1d7475ff0349f084be1d3c933c521d43f909834be4dc5d8bf7115e22cf721ba8\": container with ID starting with 1d7475ff0349f084be1d3c933c521d43f909834be4dc5d8bf7115e22cf721ba8 not found: ID does not exist" Oct 02 17:06:46 crc kubenswrapper[4808]: I1002 17:06:46.487187 4808 scope.go:117] "RemoveContainer" containerID="7dd6ccc3300b06071d3659aa79c743b63cb130b7b050cb0e88da25c34920e581" Oct 02 17:06:46 crc kubenswrapper[4808]: E1002 17:06:46.487685 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dd6ccc3300b06071d3659aa79c743b63cb130b7b050cb0e88da25c34920e581\": container with ID starting with 7dd6ccc3300b06071d3659aa79c743b63cb130b7b050cb0e88da25c34920e581 not found: ID does not exist" containerID="7dd6ccc3300b06071d3659aa79c743b63cb130b7b050cb0e88da25c34920e581" Oct 02 17:06:46 crc kubenswrapper[4808]: I1002 17:06:46.487725 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dd6ccc3300b06071d3659aa79c743b63cb130b7b050cb0e88da25c34920e581"} err="failed to get container status \"7dd6ccc3300b06071d3659aa79c743b63cb130b7b050cb0e88da25c34920e581\": rpc error: code = NotFound desc = could not find container \"7dd6ccc3300b06071d3659aa79c743b63cb130b7b050cb0e88da25c34920e581\": container with ID starting with 7dd6ccc3300b06071d3659aa79c743b63cb130b7b050cb0e88da25c34920e581 not found: ID does not exist" Oct 02 17:06:47 crc kubenswrapper[4808]: I1002 17:06:47.415447 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc0a2ce9-29ac-476e-8028-a7167f9c5ef0" path="/var/lib/kubelet/pods/dc0a2ce9-29ac-476e-8028-a7167f9c5ef0/volumes" Oct 02 17:06:57 crc kubenswrapper[4808]: I1002 17:06:57.478477 4808 generic.go:334] "Generic (PLEG): container finished" podID="fb583e21-7f4e-40c6-b028-5209db1f4603" containerID="3d3ea3025dbb30e17129dc61f4937dfd4445737ebefec1770b8ca9fe23dafff4" exitCode=0 Oct 02 17:06:57 crc kubenswrapper[4808]: I1002 17:06:57.478623 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jdsw/must-gather-22xjr" event={"ID":"fb583e21-7f4e-40c6-b028-5209db1f4603","Type":"ContainerDied","Data":"3d3ea3025dbb30e17129dc61f4937dfd4445737ebefec1770b8ca9fe23dafff4"} Oct 02 17:06:57 crc kubenswrapper[4808]: I1002 17:06:57.479919 4808 scope.go:117] "RemoveContainer" containerID="3d3ea3025dbb30e17129dc61f4937dfd4445737ebefec1770b8ca9fe23dafff4" Oct 02 17:06:58 crc kubenswrapper[4808]: I1002 17:06:58.105095 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5jdsw_must-gather-22xjr_fb583e21-7f4e-40c6-b028-5209db1f4603/gather/0.log" Oct 02 17:06:59 crc kubenswrapper[4808]: I1002 17:06:59.057928 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-6bh5l"] Oct 02 17:06:59 crc kubenswrapper[4808]: I1002 17:06:59.073950 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-rr4z9"] Oct 02 17:06:59 crc kubenswrapper[4808]: I1002 17:06:59.082987 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-t6tjw"] Oct 02 17:06:59 crc kubenswrapper[4808]: I1002 17:06:59.089943 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-6bh5l"] Oct 02 17:06:59 crc kubenswrapper[4808]: I1002 17:06:59.098255 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-t6tjw"] Oct 02 17:06:59 crc kubenswrapper[4808]: I1002 17:06:59.105548 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-rr4z9"] Oct 02 17:06:59 crc kubenswrapper[4808]: I1002 17:06:59.410712 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f95dff9-cca1-4356-912a-76d1475a273a" path="/var/lib/kubelet/pods/9f95dff9-cca1-4356-912a-76d1475a273a/volumes" Oct 02 17:06:59 crc kubenswrapper[4808]: I1002 17:06:59.411867 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af84bfb8-1583-46a8-ae8f-1b28b3568311" path="/var/lib/kubelet/pods/af84bfb8-1583-46a8-ae8f-1b28b3568311/volumes" Oct 02 17:06:59 crc kubenswrapper[4808]: I1002 17:06:59.434382 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2120107-7ced-4921-9a2d-eba436f36928" path="/var/lib/kubelet/pods/f2120107-7ced-4921-9a2d-eba436f36928/volumes" Oct 02 17:07:02 crc kubenswrapper[4808]: I1002 17:07:02.049760 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-67d46"] Oct 02 17:07:02 crc kubenswrapper[4808]: I1002 17:07:02.065570 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-67d46"] Oct 02 17:07:03 crc kubenswrapper[4808]: I1002 17:07:03.036671 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-7wz2m"] Oct 02 17:07:03 crc kubenswrapper[4808]: I1002 17:07:03.052180 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-7wz2m"] Oct 02 17:07:03 crc kubenswrapper[4808]: I1002 17:07:03.432617 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a5d0b28-6ba9-4421-98b6-1c6280ed7937" path="/var/lib/kubelet/pods/8a5d0b28-6ba9-4421-98b6-1c6280ed7937/volumes" Oct 02 17:07:03 crc kubenswrapper[4808]: I1002 17:07:03.434120 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfeb1719-f48e-488b-8c22-a9be946819e3" path="/var/lib/kubelet/pods/cfeb1719-f48e-488b-8c22-a9be946819e3/volumes" Oct 02 17:07:05 crc kubenswrapper[4808]: I1002 17:07:05.915032 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5jdsw/must-gather-22xjr"] Oct 02 17:07:05 crc kubenswrapper[4808]: I1002 17:07:05.915640 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-5jdsw/must-gather-22xjr" podUID="fb583e21-7f4e-40c6-b028-5209db1f4603" containerName="copy" containerID="cri-o://85efa474b6eaea6575403321be15c31d92778050b859a1ae7d700e35eec1ead4" gracePeriod=2 Oct 02 17:07:05 crc kubenswrapper[4808]: I1002 17:07:05.923520 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5jdsw/must-gather-22xjr"] Oct 02 17:07:06 crc kubenswrapper[4808]: I1002 17:07:06.322204 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5jdsw_must-gather-22xjr_fb583e21-7f4e-40c6-b028-5209db1f4603/copy/0.log" Oct 02 17:07:06 crc kubenswrapper[4808]: I1002 17:07:06.322917 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jdsw/must-gather-22xjr" Oct 02 17:07:06 crc kubenswrapper[4808]: I1002 17:07:06.478987 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fb583e21-7f4e-40c6-b028-5209db1f4603-must-gather-output\") pod \"fb583e21-7f4e-40c6-b028-5209db1f4603\" (UID: \"fb583e21-7f4e-40c6-b028-5209db1f4603\") " Oct 02 17:07:06 crc kubenswrapper[4808]: I1002 17:07:06.479119 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rs4w\" (UniqueName: \"kubernetes.io/projected/fb583e21-7f4e-40c6-b028-5209db1f4603-kube-api-access-2rs4w\") pod \"fb583e21-7f4e-40c6-b028-5209db1f4603\" (UID: \"fb583e21-7f4e-40c6-b028-5209db1f4603\") " Oct 02 17:07:06 crc kubenswrapper[4808]: I1002 17:07:06.486915 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb583e21-7f4e-40c6-b028-5209db1f4603-kube-api-access-2rs4w" (OuterVolumeSpecName: "kube-api-access-2rs4w") pod "fb583e21-7f4e-40c6-b028-5209db1f4603" (UID: "fb583e21-7f4e-40c6-b028-5209db1f4603"). InnerVolumeSpecName "kube-api-access-2rs4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:07:06 crc kubenswrapper[4808]: I1002 17:07:06.582158 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5jdsw_must-gather-22xjr_fb583e21-7f4e-40c6-b028-5209db1f4603/copy/0.log" Oct 02 17:07:06 crc kubenswrapper[4808]: I1002 17:07:06.582760 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rs4w\" (UniqueName: \"kubernetes.io/projected/fb583e21-7f4e-40c6-b028-5209db1f4603-kube-api-access-2rs4w\") on node \"crc\" DevicePath \"\"" Oct 02 17:07:06 crc kubenswrapper[4808]: I1002 17:07:06.582893 4808 generic.go:334] "Generic (PLEG): container finished" podID="fb583e21-7f4e-40c6-b028-5209db1f4603" containerID="85efa474b6eaea6575403321be15c31d92778050b859a1ae7d700e35eec1ead4" exitCode=143 Oct 02 17:07:06 crc kubenswrapper[4808]: I1002 17:07:06.582953 4808 scope.go:117] "RemoveContainer" containerID="85efa474b6eaea6575403321be15c31d92778050b859a1ae7d700e35eec1ead4" Oct 02 17:07:06 crc kubenswrapper[4808]: I1002 17:07:06.583078 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jdsw/must-gather-22xjr" Oct 02 17:07:06 crc kubenswrapper[4808]: I1002 17:07:06.607619 4808 scope.go:117] "RemoveContainer" containerID="3d3ea3025dbb30e17129dc61f4937dfd4445737ebefec1770b8ca9fe23dafff4" Oct 02 17:07:06 crc kubenswrapper[4808]: I1002 17:07:06.714372 4808 scope.go:117] "RemoveContainer" containerID="85efa474b6eaea6575403321be15c31d92778050b859a1ae7d700e35eec1ead4" Oct 02 17:07:06 crc kubenswrapper[4808]: E1002 17:07:06.714931 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85efa474b6eaea6575403321be15c31d92778050b859a1ae7d700e35eec1ead4\": container with ID starting with 85efa474b6eaea6575403321be15c31d92778050b859a1ae7d700e35eec1ead4 not found: ID does not exist" containerID="85efa474b6eaea6575403321be15c31d92778050b859a1ae7d700e35eec1ead4" Oct 02 17:07:06 crc kubenswrapper[4808]: I1002 17:07:06.714975 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85efa474b6eaea6575403321be15c31d92778050b859a1ae7d700e35eec1ead4"} err="failed to get container status \"85efa474b6eaea6575403321be15c31d92778050b859a1ae7d700e35eec1ead4\": rpc error: code = NotFound desc = could not find container \"85efa474b6eaea6575403321be15c31d92778050b859a1ae7d700e35eec1ead4\": container with ID starting with 85efa474b6eaea6575403321be15c31d92778050b859a1ae7d700e35eec1ead4 not found: ID does not exist" Oct 02 17:07:06 crc kubenswrapper[4808]: I1002 17:07:06.715032 4808 scope.go:117] "RemoveContainer" containerID="3d3ea3025dbb30e17129dc61f4937dfd4445737ebefec1770b8ca9fe23dafff4" Oct 02 17:07:06 crc kubenswrapper[4808]: E1002 17:07:06.715589 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d3ea3025dbb30e17129dc61f4937dfd4445737ebefec1770b8ca9fe23dafff4\": container with ID starting with 3d3ea3025dbb30e17129dc61f4937dfd4445737ebefec1770b8ca9fe23dafff4 not found: ID does not exist" containerID="3d3ea3025dbb30e17129dc61f4937dfd4445737ebefec1770b8ca9fe23dafff4" Oct 02 17:07:06 crc kubenswrapper[4808]: I1002 17:07:06.715643 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d3ea3025dbb30e17129dc61f4937dfd4445737ebefec1770b8ca9fe23dafff4"} err="failed to get container status \"3d3ea3025dbb30e17129dc61f4937dfd4445737ebefec1770b8ca9fe23dafff4\": rpc error: code = NotFound desc = could not find container \"3d3ea3025dbb30e17129dc61f4937dfd4445737ebefec1770b8ca9fe23dafff4\": container with ID starting with 3d3ea3025dbb30e17129dc61f4937dfd4445737ebefec1770b8ca9fe23dafff4 not found: ID does not exist" Oct 02 17:07:06 crc kubenswrapper[4808]: I1002 17:07:06.771725 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb583e21-7f4e-40c6-b028-5209db1f4603-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "fb583e21-7f4e-40c6-b028-5209db1f4603" (UID: "fb583e21-7f4e-40c6-b028-5209db1f4603"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 17:07:06 crc kubenswrapper[4808]: I1002 17:07:06.786586 4808 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fb583e21-7f4e-40c6-b028-5209db1f4603-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 02 17:07:07 crc kubenswrapper[4808]: I1002 17:07:07.407663 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb583e21-7f4e-40c6-b028-5209db1f4603" path="/var/lib/kubelet/pods/fb583e21-7f4e-40c6-b028-5209db1f4603/volumes" Oct 02 17:07:10 crc kubenswrapper[4808]: I1002 17:07:10.030864 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-4027-account-create-77w26"] Oct 02 17:07:10 crc kubenswrapper[4808]: I1002 17:07:10.038431 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7c90-account-create-z4xqz"] Oct 02 17:07:10 crc kubenswrapper[4808]: I1002 17:07:10.046295 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-4027-account-create-77w26"] Oct 02 17:07:10 crc kubenswrapper[4808]: I1002 17:07:10.056001 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-7b56-account-create-5z4wr"] Oct 02 17:07:10 crc kubenswrapper[4808]: I1002 17:07:10.062973 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7c90-account-create-z4xqz"] Oct 02 17:07:10 crc kubenswrapper[4808]: I1002 17:07:10.070410 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-7b56-account-create-5z4wr"] Oct 02 17:07:11 crc kubenswrapper[4808]: I1002 17:07:11.406624 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7951034c-d159-4816-8f1e-4118953ac9f0" path="/var/lib/kubelet/pods/7951034c-d159-4816-8f1e-4118953ac9f0/volumes" Oct 02 17:07:11 crc kubenswrapper[4808]: I1002 17:07:11.408181 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afc592f9-999b-4787-ae91-bdcd10aa2ad5" path="/var/lib/kubelet/pods/afc592f9-999b-4787-ae91-bdcd10aa2ad5/volumes" Oct 02 17:07:11 crc kubenswrapper[4808]: I1002 17:07:11.408830 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a" path="/var/lib/kubelet/pods/f9ad3dd3-b58c-4b3b-8b58-f0a21654ec2a/volumes" Oct 02 17:07:17 crc kubenswrapper[4808]: I1002 17:07:17.692591 4808 generic.go:334] "Generic (PLEG): container finished" podID="9a18970b-8022-4e05-8068-77b866f91b82" containerID="7e0ceabd97171ee9db645825cae3cab7ab3300b18eea10a00e19260d4972e829" exitCode=0 Oct 02 17:07:17 crc kubenswrapper[4808]: I1002 17:07:17.692672 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq" event={"ID":"9a18970b-8022-4e05-8068-77b866f91b82","Type":"ContainerDied","Data":"7e0ceabd97171ee9db645825cae3cab7ab3300b18eea10a00e19260d4972e829"} Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.188000 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.325481 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a18970b-8022-4e05-8068-77b866f91b82-ssh-key\") pod \"9a18970b-8022-4e05-8068-77b866f91b82\" (UID: \"9a18970b-8022-4e05-8068-77b866f91b82\") " Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.325570 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a18970b-8022-4e05-8068-77b866f91b82-inventory\") pod \"9a18970b-8022-4e05-8068-77b866f91b82\" (UID: \"9a18970b-8022-4e05-8068-77b866f91b82\") " Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.325755 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgnb8\" (UniqueName: \"kubernetes.io/projected/9a18970b-8022-4e05-8068-77b866f91b82-kube-api-access-sgnb8\") pod \"9a18970b-8022-4e05-8068-77b866f91b82\" (UID: \"9a18970b-8022-4e05-8068-77b866f91b82\") " Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.339479 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a18970b-8022-4e05-8068-77b866f91b82-kube-api-access-sgnb8" (OuterVolumeSpecName: "kube-api-access-sgnb8") pod "9a18970b-8022-4e05-8068-77b866f91b82" (UID: "9a18970b-8022-4e05-8068-77b866f91b82"). InnerVolumeSpecName "kube-api-access-sgnb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.353405 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a18970b-8022-4e05-8068-77b866f91b82-inventory" (OuterVolumeSpecName: "inventory") pod "9a18970b-8022-4e05-8068-77b866f91b82" (UID: "9a18970b-8022-4e05-8068-77b866f91b82"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.361611 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a18970b-8022-4e05-8068-77b866f91b82-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9a18970b-8022-4e05-8068-77b866f91b82" (UID: "9a18970b-8022-4e05-8068-77b866f91b82"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.427996 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a18970b-8022-4e05-8068-77b866f91b82-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.428042 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a18970b-8022-4e05-8068-77b866f91b82-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.428062 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgnb8\" (UniqueName: \"kubernetes.io/projected/9a18970b-8022-4e05-8068-77b866f91b82-kube-api-access-sgnb8\") on node \"crc\" DevicePath \"\"" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.719179 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq" event={"ID":"9a18970b-8022-4e05-8068-77b866f91b82","Type":"ContainerDied","Data":"8716bcfd614419032db7682e8510436135af131f8a014175f7b76dceef4bf3ce"} Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.719221 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8716bcfd614419032db7682e8510436135af131f8a014175f7b76dceef4bf3ce" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.719357 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-scfqq" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.811629 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv"] Oct 02 17:07:19 crc kubenswrapper[4808]: E1002 17:07:19.812040 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb583e21-7f4e-40c6-b028-5209db1f4603" containerName="gather" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.812060 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb583e21-7f4e-40c6-b028-5209db1f4603" containerName="gather" Oct 02 17:07:19 crc kubenswrapper[4808]: E1002 17:07:19.812074 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc0a2ce9-29ac-476e-8028-a7167f9c5ef0" containerName="extract-utilities" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.812084 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc0a2ce9-29ac-476e-8028-a7167f9c5ef0" containerName="extract-utilities" Oct 02 17:07:19 crc kubenswrapper[4808]: E1002 17:07:19.812108 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc0a2ce9-29ac-476e-8028-a7167f9c5ef0" containerName="extract-content" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.812116 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc0a2ce9-29ac-476e-8028-a7167f9c5ef0" containerName="extract-content" Oct 02 17:07:19 crc kubenswrapper[4808]: E1002 17:07:19.812130 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb583e21-7f4e-40c6-b028-5209db1f4603" containerName="copy" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.812138 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb583e21-7f4e-40c6-b028-5209db1f4603" containerName="copy" Oct 02 17:07:19 crc kubenswrapper[4808]: E1002 17:07:19.812155 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc0a2ce9-29ac-476e-8028-a7167f9c5ef0" containerName="registry-server" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.812163 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc0a2ce9-29ac-476e-8028-a7167f9c5ef0" containerName="registry-server" Oct 02 17:07:19 crc kubenswrapper[4808]: E1002 17:07:19.812181 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a18970b-8022-4e05-8068-77b866f91b82" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.812190 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a18970b-8022-4e05-8068-77b866f91b82" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.812407 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb583e21-7f4e-40c6-b028-5209db1f4603" containerName="copy" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.812432 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a18970b-8022-4e05-8068-77b866f91b82" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.812448 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc0a2ce9-29ac-476e-8028-a7167f9c5ef0" containerName="registry-server" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.812460 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb583e21-7f4e-40c6-b028-5209db1f4603" containerName="gather" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.813107 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.815770 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.816120 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.817586 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.827162 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6w9jl" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.850681 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv"] Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.965215 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06e4d2fc-7a9a-4f57-bec8-6d49771dfe48-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv\" (UID: \"06e4d2fc-7a9a-4f57-bec8-6d49771dfe48\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.965344 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r9qf\" (UniqueName: \"kubernetes.io/projected/06e4d2fc-7a9a-4f57-bec8-6d49771dfe48-kube-api-access-9r9qf\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv\" (UID: \"06e4d2fc-7a9a-4f57-bec8-6d49771dfe48\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv" Oct 02 17:07:19 crc kubenswrapper[4808]: I1002 17:07:19.965397 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06e4d2fc-7a9a-4f57-bec8-6d49771dfe48-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv\" (UID: \"06e4d2fc-7a9a-4f57-bec8-6d49771dfe48\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv" Oct 02 17:07:20 crc kubenswrapper[4808]: I1002 17:07:20.067077 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06e4d2fc-7a9a-4f57-bec8-6d49771dfe48-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv\" (UID: \"06e4d2fc-7a9a-4f57-bec8-6d49771dfe48\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv" Oct 02 17:07:20 crc kubenswrapper[4808]: I1002 17:07:20.067661 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r9qf\" (UniqueName: \"kubernetes.io/projected/06e4d2fc-7a9a-4f57-bec8-6d49771dfe48-kube-api-access-9r9qf\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv\" (UID: \"06e4d2fc-7a9a-4f57-bec8-6d49771dfe48\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv" Oct 02 17:07:20 crc kubenswrapper[4808]: I1002 17:07:20.067918 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06e4d2fc-7a9a-4f57-bec8-6d49771dfe48-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv\" (UID: \"06e4d2fc-7a9a-4f57-bec8-6d49771dfe48\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv" Oct 02 17:07:20 crc kubenswrapper[4808]: I1002 17:07:20.075787 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06e4d2fc-7a9a-4f57-bec8-6d49771dfe48-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv\" (UID: \"06e4d2fc-7a9a-4f57-bec8-6d49771dfe48\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv" Oct 02 17:07:20 crc kubenswrapper[4808]: I1002 17:07:20.076769 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06e4d2fc-7a9a-4f57-bec8-6d49771dfe48-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv\" (UID: \"06e4d2fc-7a9a-4f57-bec8-6d49771dfe48\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv" Oct 02 17:07:20 crc kubenswrapper[4808]: I1002 17:07:20.090336 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r9qf\" (UniqueName: \"kubernetes.io/projected/06e4d2fc-7a9a-4f57-bec8-6d49771dfe48-kube-api-access-9r9qf\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv\" (UID: \"06e4d2fc-7a9a-4f57-bec8-6d49771dfe48\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv" Oct 02 17:07:20 crc kubenswrapper[4808]: I1002 17:07:20.132552 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv" Oct 02 17:07:20 crc kubenswrapper[4808]: W1002 17:07:20.751287 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06e4d2fc_7a9a_4f57_bec8_6d49771dfe48.slice/crio-0134da500daabc4afe5cdbad0801b031940a8ec0022ad64838de254187174224 WatchSource:0}: Error finding container 0134da500daabc4afe5cdbad0801b031940a8ec0022ad64838de254187174224: Status 404 returned error can't find the container with id 0134da500daabc4afe5cdbad0801b031940a8ec0022ad64838de254187174224 Oct 02 17:07:20 crc kubenswrapper[4808]: I1002 17:07:20.756162 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv"] Oct 02 17:07:21 crc kubenswrapper[4808]: I1002 17:07:21.741908 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv" event={"ID":"06e4d2fc-7a9a-4f57-bec8-6d49771dfe48","Type":"ContainerStarted","Data":"d39cff55e1a9bab8c3476cc4b7a569836263156f4ada9da981cb6ac7132a7f57"} Oct 02 17:07:21 crc kubenswrapper[4808]: I1002 17:07:21.742408 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv" event={"ID":"06e4d2fc-7a9a-4f57-bec8-6d49771dfe48","Type":"ContainerStarted","Data":"0134da500daabc4afe5cdbad0801b031940a8ec0022ad64838de254187174224"} Oct 02 17:07:21 crc kubenswrapper[4808]: I1002 17:07:21.776543 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv" podStartSLOduration=2.140051415 podStartE2EDuration="2.776513499s" podCreationTimestamp="2025-10-02 17:07:19 +0000 UTC" firstStartedPulling="2025-10-02 17:07:20.756369567 +0000 UTC m=+1608.081898567" lastFinishedPulling="2025-10-02 17:07:21.392831621 +0000 UTC m=+1608.718360651" observedRunningTime="2025-10-02 17:07:21.76196381 +0000 UTC m=+1609.087492850" watchObservedRunningTime="2025-10-02 17:07:21.776513499 +0000 UTC m=+1609.102042529" Oct 02 17:07:26 crc kubenswrapper[4808]: I1002 17:07:26.801015 4808 generic.go:334] "Generic (PLEG): container finished" podID="06e4d2fc-7a9a-4f57-bec8-6d49771dfe48" containerID="d39cff55e1a9bab8c3476cc4b7a569836263156f4ada9da981cb6ac7132a7f57" exitCode=0 Oct 02 17:07:26 crc kubenswrapper[4808]: I1002 17:07:26.801306 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv" event={"ID":"06e4d2fc-7a9a-4f57-bec8-6d49771dfe48","Type":"ContainerDied","Data":"d39cff55e1a9bab8c3476cc4b7a569836263156f4ada9da981cb6ac7132a7f57"} Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.202476 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv" Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.475584 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06e4d2fc-7a9a-4f57-bec8-6d49771dfe48-inventory\") pod \"06e4d2fc-7a9a-4f57-bec8-6d49771dfe48\" (UID: \"06e4d2fc-7a9a-4f57-bec8-6d49771dfe48\") " Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.475723 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9r9qf\" (UniqueName: \"kubernetes.io/projected/06e4d2fc-7a9a-4f57-bec8-6d49771dfe48-kube-api-access-9r9qf\") pod \"06e4d2fc-7a9a-4f57-bec8-6d49771dfe48\" (UID: \"06e4d2fc-7a9a-4f57-bec8-6d49771dfe48\") " Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.475781 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06e4d2fc-7a9a-4f57-bec8-6d49771dfe48-ssh-key\") pod \"06e4d2fc-7a9a-4f57-bec8-6d49771dfe48\" (UID: \"06e4d2fc-7a9a-4f57-bec8-6d49771dfe48\") " Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.482723 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06e4d2fc-7a9a-4f57-bec8-6d49771dfe48-kube-api-access-9r9qf" (OuterVolumeSpecName: "kube-api-access-9r9qf") pod "06e4d2fc-7a9a-4f57-bec8-6d49771dfe48" (UID: "06e4d2fc-7a9a-4f57-bec8-6d49771dfe48"). InnerVolumeSpecName "kube-api-access-9r9qf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.503297 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06e4d2fc-7a9a-4f57-bec8-6d49771dfe48-inventory" (OuterVolumeSpecName: "inventory") pod "06e4d2fc-7a9a-4f57-bec8-6d49771dfe48" (UID: "06e4d2fc-7a9a-4f57-bec8-6d49771dfe48"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.512572 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06e4d2fc-7a9a-4f57-bec8-6d49771dfe48-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "06e4d2fc-7a9a-4f57-bec8-6d49771dfe48" (UID: "06e4d2fc-7a9a-4f57-bec8-6d49771dfe48"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.578271 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06e4d2fc-7a9a-4f57-bec8-6d49771dfe48-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.578300 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06e4d2fc-7a9a-4f57-bec8-6d49771dfe48-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.578312 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9r9qf\" (UniqueName: \"kubernetes.io/projected/06e4d2fc-7a9a-4f57-bec8-6d49771dfe48-kube-api-access-9r9qf\") on node \"crc\" DevicePath \"\"" Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.820905 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv" event={"ID":"06e4d2fc-7a9a-4f57-bec8-6d49771dfe48","Type":"ContainerDied","Data":"0134da500daabc4afe5cdbad0801b031940a8ec0022ad64838de254187174224"} Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.821293 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0134da500daabc4afe5cdbad0801b031940a8ec0022ad64838de254187174224" Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.821042 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-n6bmv" Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.904483 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8"] Oct 02 17:07:28 crc kubenswrapper[4808]: E1002 17:07:28.905229 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06e4d2fc-7a9a-4f57-bec8-6d49771dfe48" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.905312 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="06e4d2fc-7a9a-4f57-bec8-6d49771dfe48" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.905724 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="06e4d2fc-7a9a-4f57-bec8-6d49771dfe48" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.906923 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8" Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.908999 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.909249 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.909787 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.909809 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6w9jl" Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.913172 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8"] Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.988453 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07552c88-6521-48fc-a740-15a1a4a68dd1-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8\" (UID: \"07552c88-6521-48fc-a740-15a1a4a68dd1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8" Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.988594 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5z6v\" (UniqueName: \"kubernetes.io/projected/07552c88-6521-48fc-a740-15a1a4a68dd1-kube-api-access-r5z6v\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8\" (UID: \"07552c88-6521-48fc-a740-15a1a4a68dd1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8" Oct 02 17:07:28 crc kubenswrapper[4808]: I1002 17:07:28.988656 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07552c88-6521-48fc-a740-15a1a4a68dd1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8\" (UID: \"07552c88-6521-48fc-a740-15a1a4a68dd1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8" Oct 02 17:07:29 crc kubenswrapper[4808]: I1002 17:07:29.089565 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5z6v\" (UniqueName: \"kubernetes.io/projected/07552c88-6521-48fc-a740-15a1a4a68dd1-kube-api-access-r5z6v\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8\" (UID: \"07552c88-6521-48fc-a740-15a1a4a68dd1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8" Oct 02 17:07:29 crc kubenswrapper[4808]: I1002 17:07:29.089616 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07552c88-6521-48fc-a740-15a1a4a68dd1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8\" (UID: \"07552c88-6521-48fc-a740-15a1a4a68dd1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8" Oct 02 17:07:29 crc kubenswrapper[4808]: I1002 17:07:29.089719 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07552c88-6521-48fc-a740-15a1a4a68dd1-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8\" (UID: \"07552c88-6521-48fc-a740-15a1a4a68dd1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8" Oct 02 17:07:29 crc kubenswrapper[4808]: I1002 17:07:29.096821 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07552c88-6521-48fc-a740-15a1a4a68dd1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8\" (UID: \"07552c88-6521-48fc-a740-15a1a4a68dd1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8" Oct 02 17:07:29 crc kubenswrapper[4808]: I1002 17:07:29.096956 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07552c88-6521-48fc-a740-15a1a4a68dd1-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8\" (UID: \"07552c88-6521-48fc-a740-15a1a4a68dd1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8" Oct 02 17:07:29 crc kubenswrapper[4808]: I1002 17:07:29.125908 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5z6v\" (UniqueName: \"kubernetes.io/projected/07552c88-6521-48fc-a740-15a1a4a68dd1-kube-api-access-r5z6v\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8\" (UID: \"07552c88-6521-48fc-a740-15a1a4a68dd1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8" Oct 02 17:07:29 crc kubenswrapper[4808]: I1002 17:07:29.236416 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8" Oct 02 17:07:29 crc kubenswrapper[4808]: I1002 17:07:29.597822 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8"] Oct 02 17:07:29 crc kubenswrapper[4808]: I1002 17:07:29.834607 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8" event={"ID":"07552c88-6521-48fc-a740-15a1a4a68dd1","Type":"ContainerStarted","Data":"245fe310276f295c8aeb9878c0e9698d41e3d42dbf9fd1442e198cc532124f19"} Oct 02 17:07:30 crc kubenswrapper[4808]: I1002 17:07:30.847729 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8" event={"ID":"07552c88-6521-48fc-a740-15a1a4a68dd1","Type":"ContainerStarted","Data":"67576d1c7c529e9744f2afc4abe6590594ea2bf5a832f8e2748028199a98f4df"} Oct 02 17:07:30 crc kubenswrapper[4808]: I1002 17:07:30.871649 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8" podStartSLOduration=2.444473067 podStartE2EDuration="2.871614546s" podCreationTimestamp="2025-10-02 17:07:28 +0000 UTC" firstStartedPulling="2025-10-02 17:07:29.595405184 +0000 UTC m=+1616.920934204" lastFinishedPulling="2025-10-02 17:07:30.022546653 +0000 UTC m=+1617.348075683" observedRunningTime="2025-10-02 17:07:30.871135384 +0000 UTC m=+1618.196664404" watchObservedRunningTime="2025-10-02 17:07:30.871614546 +0000 UTC m=+1618.197143546" Oct 02 17:07:35 crc kubenswrapper[4808]: I1002 17:07:35.268924 4808 scope.go:117] "RemoveContainer" containerID="d94b8b03d2b3e323bd603df1e69eb862dc5f3a6f33e19b3c9b708b39f014b4d1" Oct 02 17:07:35 crc kubenswrapper[4808]: I1002 17:07:35.302301 4808 scope.go:117] "RemoveContainer" containerID="f1b6080bd93c4a4a6d797f30ad2dff9e3464e35423aa67e539aba687184a0847" Oct 02 17:07:35 crc kubenswrapper[4808]: I1002 17:07:35.359073 4808 scope.go:117] "RemoveContainer" containerID="26ed09459c52f39abfab79656cfd8c4d8c04fc13640a2314b929e472e033c6a3" Oct 02 17:07:35 crc kubenswrapper[4808]: I1002 17:07:35.420920 4808 scope.go:117] "RemoveContainer" containerID="ca4cc9c311e0b366cbc723acdc339555b046721505c67fe0ca3c881888b76936" Oct 02 17:07:35 crc kubenswrapper[4808]: I1002 17:07:35.472890 4808 scope.go:117] "RemoveContainer" containerID="aee639c7304a0f9c96eb2bba7121405627e0ac0f7f3c886fe1ee4d878b89acb4" Oct 02 17:07:35 crc kubenswrapper[4808]: I1002 17:07:35.502270 4808 scope.go:117] "RemoveContainer" containerID="8183cfd2823a8ad77830639cf768f14779529261544b212659991b9b6fbd628a" Oct 02 17:07:35 crc kubenswrapper[4808]: I1002 17:07:35.569250 4808 scope.go:117] "RemoveContainer" containerID="9806de53560962c1aac0dcf113979f67d8523dc5a68ce13ea383eb31316defae" Oct 02 17:07:35 crc kubenswrapper[4808]: I1002 17:07:35.589316 4808 scope.go:117] "RemoveContainer" containerID="5e7f34f16e7f252280017884205c35340aad7c1c560c31c945050aff995002d2" Oct 02 17:07:35 crc kubenswrapper[4808]: I1002 17:07:35.609159 4808 scope.go:117] "RemoveContainer" containerID="5e43cb4708cd4c2e4969ef25f0191c4b0d425d70cbc4484986bd60e8b5a6bdc8" Oct 02 17:07:35 crc kubenswrapper[4808]: I1002 17:07:35.640678 4808 scope.go:117] "RemoveContainer" containerID="0c6096da033553125ede6049f433d8b25e70a63f4c7dd2ce9f4774135f9c39ef" Oct 02 17:07:35 crc kubenswrapper[4808]: I1002 17:07:35.664188 4808 scope.go:117] "RemoveContainer" containerID="82fe42107df11b8d00d292a2debac65bb0fcd09f7ebc87e92644a02e72e11349" Oct 02 17:07:48 crc kubenswrapper[4808]: I1002 17:07:48.740100 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 17:07:48 crc kubenswrapper[4808]: I1002 17:07:48.740707 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 17:08:12 crc kubenswrapper[4808]: I1002 17:08:12.049156 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-dl6kd"] Oct 02 17:08:12 crc kubenswrapper[4808]: I1002 17:08:12.060295 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-ls5jx"] Oct 02 17:08:12 crc kubenswrapper[4808]: I1002 17:08:12.091667 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-ls5jx"] Oct 02 17:08:12 crc kubenswrapper[4808]: I1002 17:08:12.105678 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-dl6kd"] Oct 02 17:08:13 crc kubenswrapper[4808]: I1002 17:08:13.419291 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60c1cab9-fa93-4572-8eb5-f9682a047645" path="/var/lib/kubelet/pods/60c1cab9-fa93-4572-8eb5-f9682a047645/volumes" Oct 02 17:08:13 crc kubenswrapper[4808]: I1002 17:08:13.421581 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a437827-755a-4648-a40e-18a8ef89355d" path="/var/lib/kubelet/pods/7a437827-755a-4648-a40e-18a8ef89355d/volumes" Oct 02 17:08:18 crc kubenswrapper[4808]: I1002 17:08:18.739600 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 17:08:18 crc kubenswrapper[4808]: I1002 17:08:18.740306 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 17:08:19 crc kubenswrapper[4808]: I1002 17:08:19.024662 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-zpkmw"] Oct 02 17:08:19 crc kubenswrapper[4808]: I1002 17:08:19.031223 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-zpkmw"] Oct 02 17:08:19 crc kubenswrapper[4808]: I1002 17:08:19.415953 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c50129b6-292e-434b-a617-2781dac4101e" path="/var/lib/kubelet/pods/c50129b6-292e-434b-a617-2781dac4101e/volumes" Oct 02 17:08:24 crc kubenswrapper[4808]: I1002 17:08:24.054775 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-ckf4j"] Oct 02 17:08:24 crc kubenswrapper[4808]: I1002 17:08:24.061772 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-ckf4j"] Oct 02 17:08:25 crc kubenswrapper[4808]: I1002 17:08:25.417347 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74a1c6ad-0412-4c46-90ce-34869c9ca9e6" path="/var/lib/kubelet/pods/74a1c6ad-0412-4c46-90ce-34869c9ca9e6/volumes" Oct 02 17:08:26 crc kubenswrapper[4808]: I1002 17:08:26.035212 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-l2fls"] Oct 02 17:08:26 crc kubenswrapper[4808]: I1002 17:08:26.043549 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-l2fls"] Oct 02 17:08:27 crc kubenswrapper[4808]: I1002 17:08:27.407368 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3d4166a-c265-4cb4-8481-32e6b0144997" path="/var/lib/kubelet/pods/b3d4166a-c265-4cb4-8481-32e6b0144997/volumes" Oct 02 17:08:27 crc kubenswrapper[4808]: I1002 17:08:27.459575 4808 generic.go:334] "Generic (PLEG): container finished" podID="07552c88-6521-48fc-a740-15a1a4a68dd1" containerID="67576d1c7c529e9744f2afc4abe6590594ea2bf5a832f8e2748028199a98f4df" exitCode=2 Oct 02 17:08:27 crc kubenswrapper[4808]: I1002 17:08:27.459615 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8" event={"ID":"07552c88-6521-48fc-a740-15a1a4a68dd1","Type":"ContainerDied","Data":"67576d1c7c529e9744f2afc4abe6590594ea2bf5a832f8e2748028199a98f4df"} Oct 02 17:08:28 crc kubenswrapper[4808]: I1002 17:08:28.877344 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8" Oct 02 17:08:28 crc kubenswrapper[4808]: I1002 17:08:28.971522 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07552c88-6521-48fc-a740-15a1a4a68dd1-inventory\") pod \"07552c88-6521-48fc-a740-15a1a4a68dd1\" (UID: \"07552c88-6521-48fc-a740-15a1a4a68dd1\") " Oct 02 17:08:28 crc kubenswrapper[4808]: I1002 17:08:28.971600 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5z6v\" (UniqueName: \"kubernetes.io/projected/07552c88-6521-48fc-a740-15a1a4a68dd1-kube-api-access-r5z6v\") pod \"07552c88-6521-48fc-a740-15a1a4a68dd1\" (UID: \"07552c88-6521-48fc-a740-15a1a4a68dd1\") " Oct 02 17:08:28 crc kubenswrapper[4808]: I1002 17:08:28.971633 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07552c88-6521-48fc-a740-15a1a4a68dd1-ssh-key\") pod \"07552c88-6521-48fc-a740-15a1a4a68dd1\" (UID: \"07552c88-6521-48fc-a740-15a1a4a68dd1\") " Oct 02 17:08:28 crc kubenswrapper[4808]: I1002 17:08:28.978042 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07552c88-6521-48fc-a740-15a1a4a68dd1-kube-api-access-r5z6v" (OuterVolumeSpecName: "kube-api-access-r5z6v") pod "07552c88-6521-48fc-a740-15a1a4a68dd1" (UID: "07552c88-6521-48fc-a740-15a1a4a68dd1"). InnerVolumeSpecName "kube-api-access-r5z6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:08:28 crc kubenswrapper[4808]: I1002 17:08:28.998702 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07552c88-6521-48fc-a740-15a1a4a68dd1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "07552c88-6521-48fc-a740-15a1a4a68dd1" (UID: "07552c88-6521-48fc-a740-15a1a4a68dd1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:08:29 crc kubenswrapper[4808]: I1002 17:08:29.011672 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07552c88-6521-48fc-a740-15a1a4a68dd1-inventory" (OuterVolumeSpecName: "inventory") pod "07552c88-6521-48fc-a740-15a1a4a68dd1" (UID: "07552c88-6521-48fc-a740-15a1a4a68dd1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:08:29 crc kubenswrapper[4808]: I1002 17:08:29.074189 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07552c88-6521-48fc-a740-15a1a4a68dd1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 17:08:29 crc kubenswrapper[4808]: I1002 17:08:29.074343 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07552c88-6521-48fc-a740-15a1a4a68dd1-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 17:08:29 crc kubenswrapper[4808]: I1002 17:08:29.074364 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5z6v\" (UniqueName: \"kubernetes.io/projected/07552c88-6521-48fc-a740-15a1a4a68dd1-kube-api-access-r5z6v\") on node \"crc\" DevicePath \"\"" Oct 02 17:08:29 crc kubenswrapper[4808]: I1002 17:08:29.480767 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8" Oct 02 17:08:29 crc kubenswrapper[4808]: I1002 17:08:29.480942 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w4qg8" event={"ID":"07552c88-6521-48fc-a740-15a1a4a68dd1","Type":"ContainerDied","Data":"245fe310276f295c8aeb9878c0e9698d41e3d42dbf9fd1442e198cc532124f19"} Oct 02 17:08:29 crc kubenswrapper[4808]: I1002 17:08:29.481650 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="245fe310276f295c8aeb9878c0e9698d41e3d42dbf9fd1442e198cc532124f19" Oct 02 17:08:35 crc kubenswrapper[4808]: I1002 17:08:35.945472 4808 scope.go:117] "RemoveContainer" containerID="2986dd928bde6cf69270dddd5c0f22363ef81625c628a0bcbea2b8e1878e72bc" Oct 02 17:08:35 crc kubenswrapper[4808]: I1002 17:08:35.980310 4808 scope.go:117] "RemoveContainer" containerID="e5e9e7e471281420db45c65650174a226228535e2ecb32d6ba170bce5321b83c" Oct 02 17:08:36 crc kubenswrapper[4808]: I1002 17:08:36.051929 4808 scope.go:117] "RemoveContainer" containerID="e11d95d6c344e8547933c8bf98688be3e6c5e1d7a02e78850fed4d4acf0bf0ed" Oct 02 17:08:36 crc kubenswrapper[4808]: I1002 17:08:36.096980 4808 scope.go:117] "RemoveContainer" containerID="e5ffb3b4bc1236be9b7bac9a4a7c6a0bdd5c0340bfc95e1c5eeed99afeae8907" Oct 02 17:08:36 crc kubenswrapper[4808]: I1002 17:08:36.142562 4808 scope.go:117] "RemoveContainer" containerID="393a485895bf15b7c626820cf9f585356a32f02586b4781d1e7507f694b63239" Oct 02 17:08:36 crc kubenswrapper[4808]: I1002 17:08:36.168813 4808 scope.go:117] "RemoveContainer" containerID="3004ddbbb72e926e110a2e3f850d513f465f8b92dd7eb9e67b1ac75f7d9d4580" Oct 02 17:08:37 crc kubenswrapper[4808]: I1002 17:08:37.038935 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf"] Oct 02 17:08:37 crc kubenswrapper[4808]: E1002 17:08:37.039515 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07552c88-6521-48fc-a740-15a1a4a68dd1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 17:08:37 crc kubenswrapper[4808]: I1002 17:08:37.039566 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="07552c88-6521-48fc-a740-15a1a4a68dd1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 17:08:37 crc kubenswrapper[4808]: I1002 17:08:37.039842 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="07552c88-6521-48fc-a740-15a1a4a68dd1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 17:08:37 crc kubenswrapper[4808]: I1002 17:08:37.044656 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf" Oct 02 17:08:37 crc kubenswrapper[4808]: I1002 17:08:37.047139 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 17:08:37 crc kubenswrapper[4808]: I1002 17:08:37.048639 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6w9jl" Oct 02 17:08:37 crc kubenswrapper[4808]: I1002 17:08:37.049043 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 17:08:37 crc kubenswrapper[4808]: I1002 17:08:37.049831 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 17:08:37 crc kubenswrapper[4808]: I1002 17:08:37.053830 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf"] Oct 02 17:08:37 crc kubenswrapper[4808]: I1002 17:08:37.131272 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc7bfa2d-d035-4f65-b819-01c1314104dd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf\" (UID: \"cc7bfa2d-d035-4f65-b819-01c1314104dd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf" Oct 02 17:08:37 crc kubenswrapper[4808]: I1002 17:08:37.131407 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc7bfa2d-d035-4f65-b819-01c1314104dd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf\" (UID: \"cc7bfa2d-d035-4f65-b819-01c1314104dd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf" Oct 02 17:08:37 crc kubenswrapper[4808]: I1002 17:08:37.131449 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slwg6\" (UniqueName: \"kubernetes.io/projected/cc7bfa2d-d035-4f65-b819-01c1314104dd-kube-api-access-slwg6\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf\" (UID: \"cc7bfa2d-d035-4f65-b819-01c1314104dd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf" Oct 02 17:08:37 crc kubenswrapper[4808]: I1002 17:08:37.233206 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc7bfa2d-d035-4f65-b819-01c1314104dd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf\" (UID: \"cc7bfa2d-d035-4f65-b819-01c1314104dd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf" Oct 02 17:08:37 crc kubenswrapper[4808]: I1002 17:08:37.233531 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slwg6\" (UniqueName: \"kubernetes.io/projected/cc7bfa2d-d035-4f65-b819-01c1314104dd-kube-api-access-slwg6\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf\" (UID: \"cc7bfa2d-d035-4f65-b819-01c1314104dd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf" Oct 02 17:08:37 crc kubenswrapper[4808]: I1002 17:08:37.233688 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc7bfa2d-d035-4f65-b819-01c1314104dd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf\" (UID: \"cc7bfa2d-d035-4f65-b819-01c1314104dd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf" Oct 02 17:08:37 crc kubenswrapper[4808]: I1002 17:08:37.241937 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc7bfa2d-d035-4f65-b819-01c1314104dd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf\" (UID: \"cc7bfa2d-d035-4f65-b819-01c1314104dd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf" Oct 02 17:08:37 crc kubenswrapper[4808]: I1002 17:08:37.245357 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc7bfa2d-d035-4f65-b819-01c1314104dd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf\" (UID: \"cc7bfa2d-d035-4f65-b819-01c1314104dd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf" Oct 02 17:08:37 crc kubenswrapper[4808]: I1002 17:08:37.262668 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slwg6\" (UniqueName: \"kubernetes.io/projected/cc7bfa2d-d035-4f65-b819-01c1314104dd-kube-api-access-slwg6\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf\" (UID: \"cc7bfa2d-d035-4f65-b819-01c1314104dd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf" Oct 02 17:08:37 crc kubenswrapper[4808]: I1002 17:08:37.376522 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf" Oct 02 17:08:37 crc kubenswrapper[4808]: I1002 17:08:37.767444 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf"] Oct 02 17:08:38 crc kubenswrapper[4808]: I1002 17:08:38.577816 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf" event={"ID":"cc7bfa2d-d035-4f65-b819-01c1314104dd","Type":"ContainerStarted","Data":"ae43d1fbf35be3762e8ee9caf48f354ca90c5e4e4720cc9c446ebf427afb69d3"} Oct 02 17:08:40 crc kubenswrapper[4808]: I1002 17:08:40.604301 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf" event={"ID":"cc7bfa2d-d035-4f65-b819-01c1314104dd","Type":"ContainerStarted","Data":"0016e6bb25b94f7d1269fe54f942b4ef2829abea331e7b13e32644c1ec7732ea"} Oct 02 17:08:40 crc kubenswrapper[4808]: I1002 17:08:40.628881 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf" podStartSLOduration=2.084231147 podStartE2EDuration="3.62886042s" podCreationTimestamp="2025-10-02 17:08:37 +0000 UTC" firstStartedPulling="2025-10-02 17:08:37.771589596 +0000 UTC m=+1685.097118596" lastFinishedPulling="2025-10-02 17:08:39.316218869 +0000 UTC m=+1686.641747869" observedRunningTime="2025-10-02 17:08:40.625012477 +0000 UTC m=+1687.950541487" watchObservedRunningTime="2025-10-02 17:08:40.62886042 +0000 UTC m=+1687.954389430" Oct 02 17:08:48 crc kubenswrapper[4808]: I1002 17:08:48.740662 4808 patch_prober.go:28] interesting pod/machine-config-daemon-7z66r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 17:08:48 crc kubenswrapper[4808]: I1002 17:08:48.741704 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 17:08:48 crc kubenswrapper[4808]: I1002 17:08:48.741787 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" Oct 02 17:08:48 crc kubenswrapper[4808]: I1002 17:08:48.742797 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a823e8cfe9d799300a4fab46e28dedc29e9fa6954d8fd2945bf7c266df1f72ba"} pod="openshift-machine-config-operator/machine-config-daemon-7z66r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 17:08:48 crc kubenswrapper[4808]: I1002 17:08:48.742903 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" containerName="machine-config-daemon" containerID="cri-o://a823e8cfe9d799300a4fab46e28dedc29e9fa6954d8fd2945bf7c266df1f72ba" gracePeriod=600 Oct 02 17:08:48 crc kubenswrapper[4808]: E1002 17:08:48.881322 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7z66r_openshift-machine-config-operator(f83ce425-101d-4489-94a4-5c256eb29328)\"" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" Oct 02 17:08:49 crc kubenswrapper[4808]: I1002 17:08:49.707542 4808 generic.go:334] "Generic (PLEG): container finished" podID="f83ce425-101d-4489-94a4-5c256eb29328" containerID="a823e8cfe9d799300a4fab46e28dedc29e9fa6954d8fd2945bf7c266df1f72ba" exitCode=0 Oct 02 17:08:49 crc kubenswrapper[4808]: I1002 17:08:49.707651 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" event={"ID":"f83ce425-101d-4489-94a4-5c256eb29328","Type":"ContainerDied","Data":"a823e8cfe9d799300a4fab46e28dedc29e9fa6954d8fd2945bf7c266df1f72ba"} Oct 02 17:08:49 crc kubenswrapper[4808]: I1002 17:08:49.707713 4808 scope.go:117] "RemoveContainer" containerID="0fcb0d9bbc76e2f98d002f08b3996525b0e799fbb853c38130d2b118c79a3da9" Oct 02 17:08:49 crc kubenswrapper[4808]: I1002 17:08:49.709193 4808 scope.go:117] "RemoveContainer" containerID="a823e8cfe9d799300a4fab46e28dedc29e9fa6954d8fd2945bf7c266df1f72ba" Oct 02 17:08:49 crc kubenswrapper[4808]: E1002 17:08:49.709887 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7z66r_openshift-machine-config-operator(f83ce425-101d-4489-94a4-5c256eb29328)\"" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" Oct 02 17:09:00 crc kubenswrapper[4808]: I1002 17:09:00.397125 4808 scope.go:117] "RemoveContainer" containerID="a823e8cfe9d799300a4fab46e28dedc29e9fa6954d8fd2945bf7c266df1f72ba" Oct 02 17:09:00 crc kubenswrapper[4808]: E1002 17:09:00.398638 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7z66r_openshift-machine-config-operator(f83ce425-101d-4489-94a4-5c256eb29328)\"" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" Oct 02 17:09:02 crc kubenswrapper[4808]: I1002 17:09:02.055779 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-2ws9d"] Oct 02 17:09:02 crc kubenswrapper[4808]: I1002 17:09:02.069865 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-nkghg"] Oct 02 17:09:02 crc kubenswrapper[4808]: I1002 17:09:02.082284 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-x5dj9"] Oct 02 17:09:02 crc kubenswrapper[4808]: I1002 17:09:02.089948 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-nkghg"] Oct 02 17:09:02 crc kubenswrapper[4808]: I1002 17:09:02.103253 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-x5dj9"] Oct 02 17:09:02 crc kubenswrapper[4808]: I1002 17:09:02.117713 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-2ws9d"] Oct 02 17:09:03 crc kubenswrapper[4808]: I1002 17:09:03.409795 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c64446b-d171-4749-a4b6-0b71c80ce9da" path="/var/lib/kubelet/pods/1c64446b-d171-4749-a4b6-0b71c80ce9da/volumes" Oct 02 17:09:03 crc kubenswrapper[4808]: I1002 17:09:03.411026 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7921b775-ed41-47c3-b334-0345e978f9d0" path="/var/lib/kubelet/pods/7921b775-ed41-47c3-b334-0345e978f9d0/volumes" Oct 02 17:09:03 crc kubenswrapper[4808]: I1002 17:09:03.411843 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91cad4f7-82d4-4b24-b77d-4e98526c6d77" path="/var/lib/kubelet/pods/91cad4f7-82d4-4b24-b77d-4e98526c6d77/volumes" Oct 02 17:09:05 crc kubenswrapper[4808]: I1002 17:09:05.883506 4808 generic.go:334] "Generic (PLEG): container finished" podID="818e02a3-653a-4b36-a68b-9d4e2df0b71e" containerID="a02553b70014af81a0fd9e883ea96771d4b4a6fd6b02b442f8fd5258fdd307ed" exitCode=1 Oct 02 17:09:05 crc kubenswrapper[4808]: I1002 17:09:05.883840 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" event={"ID":"818e02a3-653a-4b36-a68b-9d4e2df0b71e","Type":"ContainerDied","Data":"a02553b70014af81a0fd9e883ea96771d4b4a6fd6b02b442f8fd5258fdd307ed"} Oct 02 17:09:05 crc kubenswrapper[4808]: I1002 17:09:05.883891 4808 scope.go:117] "RemoveContainer" containerID="f38c1e1eaa31905554cb4c587be088935750e18481e081f3175807eb6c29368c" Oct 02 17:09:05 crc kubenswrapper[4808]: I1002 17:09:05.885613 4808 scope.go:117] "RemoveContainer" containerID="a02553b70014af81a0fd9e883ea96771d4b4a6fd6b02b442f8fd5258fdd307ed" Oct 02 17:09:05 crc kubenswrapper[4808]: E1002 17:09:05.886003 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=manager pod=keystone-operator-controller-manager-65f4bcbb94-svkdd_openstack-operators(818e02a3-653a-4b36-a68b-9d4e2df0b71e)\"" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" podUID="818e02a3-653a-4b36-a68b-9d4e2df0b71e" Oct 02 17:09:10 crc kubenswrapper[4808]: I1002 17:09:10.703057 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 17:09:10 crc kubenswrapper[4808]: I1002 17:09:10.704500 4808 scope.go:117] "RemoveContainer" containerID="a02553b70014af81a0fd9e883ea96771d4b4a6fd6b02b442f8fd5258fdd307ed" Oct 02 17:09:10 crc kubenswrapper[4808]: E1002 17:09:10.704855 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=manager pod=keystone-operator-controller-manager-65f4bcbb94-svkdd_openstack-operators(818e02a3-653a-4b36-a68b-9d4e2df0b71e)\"" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" podUID="818e02a3-653a-4b36-a68b-9d4e2df0b71e" Oct 02 17:09:12 crc kubenswrapper[4808]: I1002 17:09:12.039801 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-4976-account-create-lskbv"] Oct 02 17:09:12 crc kubenswrapper[4808]: I1002 17:09:12.054898 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-5e0c-account-create-75q46"] Oct 02 17:09:12 crc kubenswrapper[4808]: I1002 17:09:12.062452 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-731b-account-create-zrkfk"] Oct 02 17:09:12 crc kubenswrapper[4808]: I1002 17:09:12.069990 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-4976-account-create-lskbv"] Oct 02 17:09:12 crc kubenswrapper[4808]: I1002 17:09:12.080477 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-731b-account-create-zrkfk"] Oct 02 17:09:12 crc kubenswrapper[4808]: I1002 17:09:12.086734 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-5e0c-account-create-75q46"] Oct 02 17:09:13 crc kubenswrapper[4808]: I1002 17:09:13.417316 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a2cb272-0bff-49dc-bf02-114dcfda4aea" path="/var/lib/kubelet/pods/5a2cb272-0bff-49dc-bf02-114dcfda4aea/volumes" Oct 02 17:09:13 crc kubenswrapper[4808]: I1002 17:09:13.418569 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8311380-1e59-4d6c-b48f-c7dac382f325" path="/var/lib/kubelet/pods/b8311380-1e59-4d6c-b48f-c7dac382f325/volumes" Oct 02 17:09:13 crc kubenswrapper[4808]: I1002 17:09:13.419661 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e94ceeb9-6ff7-4fcd-8727-64bb034bef2f" path="/var/lib/kubelet/pods/e94ceeb9-6ff7-4fcd-8727-64bb034bef2f/volumes" Oct 02 17:09:15 crc kubenswrapper[4808]: I1002 17:09:15.397063 4808 scope.go:117] "RemoveContainer" containerID="a823e8cfe9d799300a4fab46e28dedc29e9fa6954d8fd2945bf7c266df1f72ba" Oct 02 17:09:15 crc kubenswrapper[4808]: E1002 17:09:15.397635 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7z66r_openshift-machine-config-operator(f83ce425-101d-4489-94a4-5c256eb29328)\"" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" Oct 02 17:09:20 crc kubenswrapper[4808]: I1002 17:09:20.703827 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" Oct 02 17:09:20 crc kubenswrapper[4808]: I1002 17:09:20.705354 4808 scope.go:117] "RemoveContainer" containerID="a02553b70014af81a0fd9e883ea96771d4b4a6fd6b02b442f8fd5258fdd307ed" Oct 02 17:09:20 crc kubenswrapper[4808]: E1002 17:09:20.705745 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=manager pod=keystone-operator-controller-manager-65f4bcbb94-svkdd_openstack-operators(818e02a3-653a-4b36-a68b-9d4e2df0b71e)\"" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" podUID="818e02a3-653a-4b36-a68b-9d4e2df0b71e" Oct 02 17:09:28 crc kubenswrapper[4808]: I1002 17:09:28.397865 4808 scope.go:117] "RemoveContainer" containerID="a823e8cfe9d799300a4fab46e28dedc29e9fa6954d8fd2945bf7c266df1f72ba" Oct 02 17:09:28 crc kubenswrapper[4808]: E1002 17:09:28.398591 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7z66r_openshift-machine-config-operator(f83ce425-101d-4489-94a4-5c256eb29328)\"" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" Oct 02 17:09:31 crc kubenswrapper[4808]: I1002 17:09:31.178178 4808 generic.go:334] "Generic (PLEG): container finished" podID="cc7bfa2d-d035-4f65-b819-01c1314104dd" containerID="0016e6bb25b94f7d1269fe54f942b4ef2829abea331e7b13e32644c1ec7732ea" exitCode=0 Oct 02 17:09:31 crc kubenswrapper[4808]: I1002 17:09:31.178324 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf" event={"ID":"cc7bfa2d-d035-4f65-b819-01c1314104dd","Type":"ContainerDied","Data":"0016e6bb25b94f7d1269fe54f942b4ef2829abea331e7b13e32644c1ec7732ea"} Oct 02 17:09:31 crc kubenswrapper[4808]: I1002 17:09:31.396739 4808 scope.go:117] "RemoveContainer" containerID="a02553b70014af81a0fd9e883ea96771d4b4a6fd6b02b442f8fd5258fdd307ed" Oct 02 17:09:31 crc kubenswrapper[4808]: E1002 17:09:31.397015 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=manager pod=keystone-operator-controller-manager-65f4bcbb94-svkdd_openstack-operators(818e02a3-653a-4b36-a68b-9d4e2df0b71e)\"" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" podUID="818e02a3-653a-4b36-a68b-9d4e2df0b71e" Oct 02 17:09:32 crc kubenswrapper[4808]: I1002 17:09:32.056628 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hmw58"] Oct 02 17:09:32 crc kubenswrapper[4808]: I1002 17:09:32.064025 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hmw58"] Oct 02 17:09:32 crc kubenswrapper[4808]: I1002 17:09:32.590902 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf" Oct 02 17:09:32 crc kubenswrapper[4808]: I1002 17:09:32.754693 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc7bfa2d-d035-4f65-b819-01c1314104dd-inventory\") pod \"cc7bfa2d-d035-4f65-b819-01c1314104dd\" (UID: \"cc7bfa2d-d035-4f65-b819-01c1314104dd\") " Oct 02 17:09:32 crc kubenswrapper[4808]: I1002 17:09:32.754810 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc7bfa2d-d035-4f65-b819-01c1314104dd-ssh-key\") pod \"cc7bfa2d-d035-4f65-b819-01c1314104dd\" (UID: \"cc7bfa2d-d035-4f65-b819-01c1314104dd\") " Oct 02 17:09:32 crc kubenswrapper[4808]: I1002 17:09:32.754893 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slwg6\" (UniqueName: \"kubernetes.io/projected/cc7bfa2d-d035-4f65-b819-01c1314104dd-kube-api-access-slwg6\") pod \"cc7bfa2d-d035-4f65-b819-01c1314104dd\" (UID: \"cc7bfa2d-d035-4f65-b819-01c1314104dd\") " Oct 02 17:09:32 crc kubenswrapper[4808]: I1002 17:09:32.761470 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc7bfa2d-d035-4f65-b819-01c1314104dd-kube-api-access-slwg6" (OuterVolumeSpecName: "kube-api-access-slwg6") pod "cc7bfa2d-d035-4f65-b819-01c1314104dd" (UID: "cc7bfa2d-d035-4f65-b819-01c1314104dd"). InnerVolumeSpecName "kube-api-access-slwg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:09:32 crc kubenswrapper[4808]: I1002 17:09:32.785376 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc7bfa2d-d035-4f65-b819-01c1314104dd-inventory" (OuterVolumeSpecName: "inventory") pod "cc7bfa2d-d035-4f65-b819-01c1314104dd" (UID: "cc7bfa2d-d035-4f65-b819-01c1314104dd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:09:32 crc kubenswrapper[4808]: I1002 17:09:32.794375 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc7bfa2d-d035-4f65-b819-01c1314104dd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cc7bfa2d-d035-4f65-b819-01c1314104dd" (UID: "cc7bfa2d-d035-4f65-b819-01c1314104dd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:09:32 crc kubenswrapper[4808]: I1002 17:09:32.857149 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc7bfa2d-d035-4f65-b819-01c1314104dd-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 17:09:32 crc kubenswrapper[4808]: I1002 17:09:32.857201 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc7bfa2d-d035-4f65-b819-01c1314104dd-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 17:09:32 crc kubenswrapper[4808]: I1002 17:09:32.857219 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slwg6\" (UniqueName: \"kubernetes.io/projected/cc7bfa2d-d035-4f65-b819-01c1314104dd-kube-api-access-slwg6\") on node \"crc\" DevicePath \"\"" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.200471 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf" event={"ID":"cc7bfa2d-d035-4f65-b819-01c1314104dd","Type":"ContainerDied","Data":"ae43d1fbf35be3762e8ee9caf48f354ca90c5e4e4720cc9c446ebf427afb69d3"} Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.200768 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae43d1fbf35be3762e8ee9caf48f354ca90c5e4e4720cc9c446ebf427afb69d3" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.200548 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xxsf" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.291494 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-wr8qt"] Oct 02 17:09:33 crc kubenswrapper[4808]: E1002 17:09:33.292049 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc7bfa2d-d035-4f65-b819-01c1314104dd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.292080 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc7bfa2d-d035-4f65-b819-01c1314104dd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.292448 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc7bfa2d-d035-4f65-b819-01c1314104dd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.293438 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-wr8qt" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.299565 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-wr8qt"] Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.301698 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.302308 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6w9jl" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.302698 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.302879 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.366346 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/13ed2118-981b-4918-b169-0631dd74d8c9-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-wr8qt\" (UID: \"13ed2118-981b-4918-b169-0631dd74d8c9\") " pod="openstack/ssh-known-hosts-edpm-deployment-wr8qt" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.366532 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf77s\" (UniqueName: \"kubernetes.io/projected/13ed2118-981b-4918-b169-0631dd74d8c9-kube-api-access-qf77s\") pod \"ssh-known-hosts-edpm-deployment-wr8qt\" (UID: \"13ed2118-981b-4918-b169-0631dd74d8c9\") " pod="openstack/ssh-known-hosts-edpm-deployment-wr8qt" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.366570 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/13ed2118-981b-4918-b169-0631dd74d8c9-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-wr8qt\" (UID: \"13ed2118-981b-4918-b169-0631dd74d8c9\") " pod="openstack/ssh-known-hosts-edpm-deployment-wr8qt" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.425354 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81cc1bfd-c3b7-49e1-9bc7-466483bde77b" path="/var/lib/kubelet/pods/81cc1bfd-c3b7-49e1-9bc7-466483bde77b/volumes" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.469386 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf77s\" (UniqueName: \"kubernetes.io/projected/13ed2118-981b-4918-b169-0631dd74d8c9-kube-api-access-qf77s\") pod \"ssh-known-hosts-edpm-deployment-wr8qt\" (UID: \"13ed2118-981b-4918-b169-0631dd74d8c9\") " pod="openstack/ssh-known-hosts-edpm-deployment-wr8qt" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.469464 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/13ed2118-981b-4918-b169-0631dd74d8c9-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-wr8qt\" (UID: \"13ed2118-981b-4918-b169-0631dd74d8c9\") " pod="openstack/ssh-known-hosts-edpm-deployment-wr8qt" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.469512 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/13ed2118-981b-4918-b169-0631dd74d8c9-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-wr8qt\" (UID: \"13ed2118-981b-4918-b169-0631dd74d8c9\") " pod="openstack/ssh-known-hosts-edpm-deployment-wr8qt" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.474781 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.475077 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.495260 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/13ed2118-981b-4918-b169-0631dd74d8c9-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-wr8qt\" (UID: \"13ed2118-981b-4918-b169-0631dd74d8c9\") " pod="openstack/ssh-known-hosts-edpm-deployment-wr8qt" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.503212 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/13ed2118-981b-4918-b169-0631dd74d8c9-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-wr8qt\" (UID: \"13ed2118-981b-4918-b169-0631dd74d8c9\") " pod="openstack/ssh-known-hosts-edpm-deployment-wr8qt" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.523942 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf77s\" (UniqueName: \"kubernetes.io/projected/13ed2118-981b-4918-b169-0631dd74d8c9-kube-api-access-qf77s\") pod \"ssh-known-hosts-edpm-deployment-wr8qt\" (UID: \"13ed2118-981b-4918-b169-0631dd74d8c9\") " pod="openstack/ssh-known-hosts-edpm-deployment-wr8qt" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.614977 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6w9jl" Oct 02 17:09:33 crc kubenswrapper[4808]: I1002 17:09:33.623379 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-wr8qt" Oct 02 17:09:34 crc kubenswrapper[4808]: I1002 17:09:34.135372 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-wr8qt"] Oct 02 17:09:34 crc kubenswrapper[4808]: I1002 17:09:34.208839 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-wr8qt" event={"ID":"13ed2118-981b-4918-b169-0631dd74d8c9","Type":"ContainerStarted","Data":"67e4ebcc4e005953fc9694d0225462a65c10e4fa1aaa176338322064ecfcebaf"} Oct 02 17:09:34 crc kubenswrapper[4808]: I1002 17:09:34.682693 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 17:09:35 crc kubenswrapper[4808]: I1002 17:09:35.223028 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-wr8qt" event={"ID":"13ed2118-981b-4918-b169-0631dd74d8c9","Type":"ContainerStarted","Data":"ba478c0e3a11d4cb65c065bc4abec797c026e8152ca61109c8a0e1dc0867fb31"} Oct 02 17:09:35 crc kubenswrapper[4808]: I1002 17:09:35.253994 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-wr8qt" podStartSLOduration=1.732406692 podStartE2EDuration="2.253959517s" podCreationTimestamp="2025-10-02 17:09:33 +0000 UTC" firstStartedPulling="2025-10-02 17:09:34.15747644 +0000 UTC m=+1741.483005460" lastFinishedPulling="2025-10-02 17:09:34.679029245 +0000 UTC m=+1742.004558285" observedRunningTime="2025-10-02 17:09:35.242292965 +0000 UTC m=+1742.567822005" watchObservedRunningTime="2025-10-02 17:09:35.253959517 +0000 UTC m=+1742.579488557" Oct 02 17:09:36 crc kubenswrapper[4808]: I1002 17:09:36.312450 4808 scope.go:117] "RemoveContainer" containerID="e1c9417e4f2dc43ca95e4a3755f8d9316ee5b61c0677dda5c6e20ebeb9efdbc3" Oct 02 17:09:36 crc kubenswrapper[4808]: I1002 17:09:36.340379 4808 scope.go:117] "RemoveContainer" containerID="32750afea91cfd59fed83d3df033a4c648e3f9ac68afbd02f02fef845fc8251e" Oct 02 17:09:36 crc kubenswrapper[4808]: I1002 17:09:36.400017 4808 scope.go:117] "RemoveContainer" containerID="9febe6f90b26edcc63c1ec945e665d6ef621b9b737b6428b9e16ac75e6bc8dba" Oct 02 17:09:36 crc kubenswrapper[4808]: I1002 17:09:36.452618 4808 scope.go:117] "RemoveContainer" containerID="047213d1e0e0840b8ac9ea10d6144e60fbe9035487013711178ba6927959acad" Oct 02 17:09:36 crc kubenswrapper[4808]: I1002 17:09:36.477871 4808 scope.go:117] "RemoveContainer" containerID="af9a06024f8a93be5e96603b479ae8bae2cafda0340872e07d2f8d59b2768dbd" Oct 02 17:09:36 crc kubenswrapper[4808]: I1002 17:09:36.534658 4808 scope.go:117] "RemoveContainer" containerID="fa1dcef324d54debc7e14b27ead8ec6020040e475f714145f1b3489b17524f76" Oct 02 17:09:36 crc kubenswrapper[4808]: I1002 17:09:36.556184 4808 scope.go:117] "RemoveContainer" containerID="bd67e11c8060c1fe134a3721e2c51c04da63cbb6ea023b09aeaf6b6ea1b10f3a" Oct 02 17:09:40 crc kubenswrapper[4808]: I1002 17:09:40.396368 4808 scope.go:117] "RemoveContainer" containerID="a823e8cfe9d799300a4fab46e28dedc29e9fa6954d8fd2945bf7c266df1f72ba" Oct 02 17:09:40 crc kubenswrapper[4808]: E1002 17:09:40.396708 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7z66r_openshift-machine-config-operator(f83ce425-101d-4489-94a4-5c256eb29328)\"" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" Oct 02 17:09:43 crc kubenswrapper[4808]: I1002 17:09:43.306225 4808 generic.go:334] "Generic (PLEG): container finished" podID="13ed2118-981b-4918-b169-0631dd74d8c9" containerID="ba478c0e3a11d4cb65c065bc4abec797c026e8152ca61109c8a0e1dc0867fb31" exitCode=0 Oct 02 17:09:43 crc kubenswrapper[4808]: I1002 17:09:43.306331 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-wr8qt" event={"ID":"13ed2118-981b-4918-b169-0631dd74d8c9","Type":"ContainerDied","Data":"ba478c0e3a11d4cb65c065bc4abec797c026e8152ca61109c8a0e1dc0867fb31"} Oct 02 17:09:43 crc kubenswrapper[4808]: I1002 17:09:43.408389 4808 scope.go:117] "RemoveContainer" containerID="a02553b70014af81a0fd9e883ea96771d4b4a6fd6b02b442f8fd5258fdd307ed" Oct 02 17:09:43 crc kubenswrapper[4808]: E1002 17:09:43.409115 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=manager pod=keystone-operator-controller-manager-65f4bcbb94-svkdd_openstack-operators(818e02a3-653a-4b36-a68b-9d4e2df0b71e)\"" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" podUID="818e02a3-653a-4b36-a68b-9d4e2df0b71e" Oct 02 17:09:44 crc kubenswrapper[4808]: I1002 17:09:44.758411 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-wr8qt" Oct 02 17:09:44 crc kubenswrapper[4808]: I1002 17:09:44.898592 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/13ed2118-981b-4918-b169-0631dd74d8c9-ssh-key-openstack-edpm-ipam\") pod \"13ed2118-981b-4918-b169-0631dd74d8c9\" (UID: \"13ed2118-981b-4918-b169-0631dd74d8c9\") " Oct 02 17:09:44 crc kubenswrapper[4808]: I1002 17:09:44.898690 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qf77s\" (UniqueName: \"kubernetes.io/projected/13ed2118-981b-4918-b169-0631dd74d8c9-kube-api-access-qf77s\") pod \"13ed2118-981b-4918-b169-0631dd74d8c9\" (UID: \"13ed2118-981b-4918-b169-0631dd74d8c9\") " Oct 02 17:09:44 crc kubenswrapper[4808]: I1002 17:09:44.898841 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/13ed2118-981b-4918-b169-0631dd74d8c9-inventory-0\") pod \"13ed2118-981b-4918-b169-0631dd74d8c9\" (UID: \"13ed2118-981b-4918-b169-0631dd74d8c9\") " Oct 02 17:09:44 crc kubenswrapper[4808]: I1002 17:09:44.907874 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13ed2118-981b-4918-b169-0631dd74d8c9-kube-api-access-qf77s" (OuterVolumeSpecName: "kube-api-access-qf77s") pod "13ed2118-981b-4918-b169-0631dd74d8c9" (UID: "13ed2118-981b-4918-b169-0631dd74d8c9"). InnerVolumeSpecName "kube-api-access-qf77s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:09:44 crc kubenswrapper[4808]: I1002 17:09:44.937746 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13ed2118-981b-4918-b169-0631dd74d8c9-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "13ed2118-981b-4918-b169-0631dd74d8c9" (UID: "13ed2118-981b-4918-b169-0631dd74d8c9"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:09:44 crc kubenswrapper[4808]: I1002 17:09:44.940547 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13ed2118-981b-4918-b169-0631dd74d8c9-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "13ed2118-981b-4918-b169-0631dd74d8c9" (UID: "13ed2118-981b-4918-b169-0631dd74d8c9"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.001438 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/13ed2118-981b-4918-b169-0631dd74d8c9-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.001471 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qf77s\" (UniqueName: \"kubernetes.io/projected/13ed2118-981b-4918-b169-0631dd74d8c9-kube-api-access-qf77s\") on node \"crc\" DevicePath \"\"" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.001480 4808 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/13ed2118-981b-4918-b169-0631dd74d8c9-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.328596 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-wr8qt" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.329207 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-wr8qt" event={"ID":"13ed2118-981b-4918-b169-0631dd74d8c9","Type":"ContainerDied","Data":"67e4ebcc4e005953fc9694d0225462a65c10e4fa1aaa176338322064ecfcebaf"} Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.329303 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67e4ebcc4e005953fc9694d0225462a65c10e4fa1aaa176338322064ecfcebaf" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.424619 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t"] Oct 02 17:09:45 crc kubenswrapper[4808]: E1002 17:09:45.425255 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13ed2118-981b-4918-b169-0631dd74d8c9" containerName="ssh-known-hosts-edpm-deployment" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.425273 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="13ed2118-981b-4918-b169-0631dd74d8c9" containerName="ssh-known-hosts-edpm-deployment" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.425460 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="13ed2118-981b-4918-b169-0631dd74d8c9" containerName="ssh-known-hosts-edpm-deployment" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.426092 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.428266 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.428301 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.428306 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.428741 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6w9jl" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.443200 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t"] Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.613527 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rhdg\" (UniqueName: \"kubernetes.io/projected/7731cf7c-fc65-4208-9a2f-0610c7c9e0c3-kube-api-access-8rhdg\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ffs4t\" (UID: \"7731cf7c-fc65-4208-9a2f-0610c7c9e0c3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.613800 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7731cf7c-fc65-4208-9a2f-0610c7c9e0c3-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ffs4t\" (UID: \"7731cf7c-fc65-4208-9a2f-0610c7c9e0c3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.613898 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7731cf7c-fc65-4208-9a2f-0610c7c9e0c3-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ffs4t\" (UID: \"7731cf7c-fc65-4208-9a2f-0610c7c9e0c3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.715828 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rhdg\" (UniqueName: \"kubernetes.io/projected/7731cf7c-fc65-4208-9a2f-0610c7c9e0c3-kube-api-access-8rhdg\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ffs4t\" (UID: \"7731cf7c-fc65-4208-9a2f-0610c7c9e0c3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.716065 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7731cf7c-fc65-4208-9a2f-0610c7c9e0c3-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ffs4t\" (UID: \"7731cf7c-fc65-4208-9a2f-0610c7c9e0c3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.716150 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7731cf7c-fc65-4208-9a2f-0610c7c9e0c3-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ffs4t\" (UID: \"7731cf7c-fc65-4208-9a2f-0610c7c9e0c3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.732173 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7731cf7c-fc65-4208-9a2f-0610c7c9e0c3-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ffs4t\" (UID: \"7731cf7c-fc65-4208-9a2f-0610c7c9e0c3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.736258 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7731cf7c-fc65-4208-9a2f-0610c7c9e0c3-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ffs4t\" (UID: \"7731cf7c-fc65-4208-9a2f-0610c7c9e0c3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t" Oct 02 17:09:45 crc kubenswrapper[4808]: I1002 17:09:45.748683 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rhdg\" (UniqueName: \"kubernetes.io/projected/7731cf7c-fc65-4208-9a2f-0610c7c9e0c3-kube-api-access-8rhdg\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ffs4t\" (UID: \"7731cf7c-fc65-4208-9a2f-0610c7c9e0c3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t" Oct 02 17:09:46 crc kubenswrapper[4808]: I1002 17:09:46.047315 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t" Oct 02 17:09:46 crc kubenswrapper[4808]: I1002 17:09:46.723967 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t"] Oct 02 17:09:47 crc kubenswrapper[4808]: I1002 17:09:47.348153 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t" event={"ID":"7731cf7c-fc65-4208-9a2f-0610c7c9e0c3","Type":"ContainerStarted","Data":"2702c535b7cace3177f2a5697f17affb55c953733029eeb9dd2cac921135dc22"} Oct 02 17:09:48 crc kubenswrapper[4808]: I1002 17:09:48.362514 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t" event={"ID":"7731cf7c-fc65-4208-9a2f-0610c7c9e0c3","Type":"ContainerStarted","Data":"7acb1abb38802a367fd19363d4d57353f24cd0e7adea8e74a719b832e94318ac"} Oct 02 17:09:48 crc kubenswrapper[4808]: I1002 17:09:48.384421 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t" podStartSLOduration=2.915125148 podStartE2EDuration="3.384395536s" podCreationTimestamp="2025-10-02 17:09:45 +0000 UTC" firstStartedPulling="2025-10-02 17:09:46.721041935 +0000 UTC m=+1754.046570975" lastFinishedPulling="2025-10-02 17:09:47.190312323 +0000 UTC m=+1754.515841363" observedRunningTime="2025-10-02 17:09:48.378681473 +0000 UTC m=+1755.704210503" watchObservedRunningTime="2025-10-02 17:09:48.384395536 +0000 UTC m=+1755.709924566" Oct 02 17:09:53 crc kubenswrapper[4808]: I1002 17:09:53.404983 4808 scope.go:117] "RemoveContainer" containerID="a823e8cfe9d799300a4fab46e28dedc29e9fa6954d8fd2945bf7c266df1f72ba" Oct 02 17:09:53 crc kubenswrapper[4808]: E1002 17:09:53.406147 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7z66r_openshift-machine-config-operator(f83ce425-101d-4489-94a4-5c256eb29328)\"" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" Oct 02 17:09:54 crc kubenswrapper[4808]: I1002 17:09:54.057024 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-x8vmm"] Oct 02 17:09:54 crc kubenswrapper[4808]: I1002 17:09:54.066604 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-x8vmm"] Oct 02 17:09:55 crc kubenswrapper[4808]: I1002 17:09:55.051365 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z46bq"] Oct 02 17:09:55 crc kubenswrapper[4808]: I1002 17:09:55.063587 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z46bq"] Oct 02 17:09:55 crc kubenswrapper[4808]: I1002 17:09:55.411578 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4bb01f7-7c71-4b46-b45d-49d3cd6b0958" path="/var/lib/kubelet/pods/c4bb01f7-7c71-4b46-b45d-49d3cd6b0958/volumes" Oct 02 17:09:55 crc kubenswrapper[4808]: I1002 17:09:55.412247 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4c1d967-acf9-4dbb-83db-1f654d4be6d2" path="/var/lib/kubelet/pods/d4c1d967-acf9-4dbb-83db-1f654d4be6d2/volumes" Oct 02 17:09:56 crc kubenswrapper[4808]: I1002 17:09:56.448786 4808 generic.go:334] "Generic (PLEG): container finished" podID="7731cf7c-fc65-4208-9a2f-0610c7c9e0c3" containerID="7acb1abb38802a367fd19363d4d57353f24cd0e7adea8e74a719b832e94318ac" exitCode=0 Oct 02 17:09:56 crc kubenswrapper[4808]: I1002 17:09:56.449184 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t" event={"ID":"7731cf7c-fc65-4208-9a2f-0610c7c9e0c3","Type":"ContainerDied","Data":"7acb1abb38802a367fd19363d4d57353f24cd0e7adea8e74a719b832e94318ac"} Oct 02 17:09:57 crc kubenswrapper[4808]: I1002 17:09:57.396515 4808 scope.go:117] "RemoveContainer" containerID="a02553b70014af81a0fd9e883ea96771d4b4a6fd6b02b442f8fd5258fdd307ed" Oct 02 17:09:57 crc kubenswrapper[4808]: E1002 17:09:57.396810 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=manager pod=keystone-operator-controller-manager-65f4bcbb94-svkdd_openstack-operators(818e02a3-653a-4b36-a68b-9d4e2df0b71e)\"" pod="openstack-operators/keystone-operator-controller-manager-65f4bcbb94-svkdd" podUID="818e02a3-653a-4b36-a68b-9d4e2df0b71e" Oct 02 17:09:57 crc kubenswrapper[4808]: I1002 17:09:57.861633 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t" Oct 02 17:09:57 crc kubenswrapper[4808]: I1002 17:09:57.977194 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7731cf7c-fc65-4208-9a2f-0610c7c9e0c3-inventory\") pod \"7731cf7c-fc65-4208-9a2f-0610c7c9e0c3\" (UID: \"7731cf7c-fc65-4208-9a2f-0610c7c9e0c3\") " Oct 02 17:09:57 crc kubenswrapper[4808]: I1002 17:09:57.977291 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7731cf7c-fc65-4208-9a2f-0610c7c9e0c3-ssh-key\") pod \"7731cf7c-fc65-4208-9a2f-0610c7c9e0c3\" (UID: \"7731cf7c-fc65-4208-9a2f-0610c7c9e0c3\") " Oct 02 17:09:57 crc kubenswrapper[4808]: I1002 17:09:57.977434 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rhdg\" (UniqueName: \"kubernetes.io/projected/7731cf7c-fc65-4208-9a2f-0610c7c9e0c3-kube-api-access-8rhdg\") pod \"7731cf7c-fc65-4208-9a2f-0610c7c9e0c3\" (UID: \"7731cf7c-fc65-4208-9a2f-0610c7c9e0c3\") " Oct 02 17:09:57 crc kubenswrapper[4808]: I1002 17:09:57.984734 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7731cf7c-fc65-4208-9a2f-0610c7c9e0c3-kube-api-access-8rhdg" (OuterVolumeSpecName: "kube-api-access-8rhdg") pod "7731cf7c-fc65-4208-9a2f-0610c7c9e0c3" (UID: "7731cf7c-fc65-4208-9a2f-0610c7c9e0c3"). InnerVolumeSpecName "kube-api-access-8rhdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.024072 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7731cf7c-fc65-4208-9a2f-0610c7c9e0c3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7731cf7c-fc65-4208-9a2f-0610c7c9e0c3" (UID: "7731cf7c-fc65-4208-9a2f-0610c7c9e0c3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.024418 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7731cf7c-fc65-4208-9a2f-0610c7c9e0c3-inventory" (OuterVolumeSpecName: "inventory") pod "7731cf7c-fc65-4208-9a2f-0610c7c9e0c3" (UID: "7731cf7c-fc65-4208-9a2f-0610c7c9e0c3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.080193 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rhdg\" (UniqueName: \"kubernetes.io/projected/7731cf7c-fc65-4208-9a2f-0610c7c9e0c3-kube-api-access-8rhdg\") on node \"crc\" DevicePath \"\"" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.080284 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7731cf7c-fc65-4208-9a2f-0610c7c9e0c3-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.080306 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7731cf7c-fc65-4208-9a2f-0610c7c9e0c3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.466487 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t" event={"ID":"7731cf7c-fc65-4208-9a2f-0610c7c9e0c3","Type":"ContainerDied","Data":"2702c535b7cace3177f2a5697f17affb55c953733029eeb9dd2cac921135dc22"} Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.466797 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2702c535b7cace3177f2a5697f17affb55c953733029eeb9dd2cac921135dc22" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.466618 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ffs4t" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.547394 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb"] Oct 02 17:09:58 crc kubenswrapper[4808]: E1002 17:09:58.547782 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7731cf7c-fc65-4208-9a2f-0610c7c9e0c3" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.547800 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7731cf7c-fc65-4208-9a2f-0610c7c9e0c3" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.548011 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7731cf7c-fc65-4208-9a2f-0610c7c9e0c3" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.548666 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.551253 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.551625 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.551788 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.557756 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6w9jl" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.561914 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb"] Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.690398 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d9cd742-d1a8-48ae-bd84-3660eb151623-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb\" (UID: \"7d9cd742-d1a8-48ae-bd84-3660eb151623\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.690573 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nks6c\" (UniqueName: \"kubernetes.io/projected/7d9cd742-d1a8-48ae-bd84-3660eb151623-kube-api-access-nks6c\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb\" (UID: \"7d9cd742-d1a8-48ae-bd84-3660eb151623\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.690603 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d9cd742-d1a8-48ae-bd84-3660eb151623-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb\" (UID: \"7d9cd742-d1a8-48ae-bd84-3660eb151623\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.793029 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d9cd742-d1a8-48ae-bd84-3660eb151623-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb\" (UID: \"7d9cd742-d1a8-48ae-bd84-3660eb151623\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.793105 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nks6c\" (UniqueName: \"kubernetes.io/projected/7d9cd742-d1a8-48ae-bd84-3660eb151623-kube-api-access-nks6c\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb\" (UID: \"7d9cd742-d1a8-48ae-bd84-3660eb151623\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.793129 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d9cd742-d1a8-48ae-bd84-3660eb151623-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb\" (UID: \"7d9cd742-d1a8-48ae-bd84-3660eb151623\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.800916 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d9cd742-d1a8-48ae-bd84-3660eb151623-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb\" (UID: \"7d9cd742-d1a8-48ae-bd84-3660eb151623\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.816194 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d9cd742-d1a8-48ae-bd84-3660eb151623-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb\" (UID: \"7d9cd742-d1a8-48ae-bd84-3660eb151623\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.818927 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nks6c\" (UniqueName: \"kubernetes.io/projected/7d9cd742-d1a8-48ae-bd84-3660eb151623-kube-api-access-nks6c\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb\" (UID: \"7d9cd742-d1a8-48ae-bd84-3660eb151623\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb" Oct 02 17:09:58 crc kubenswrapper[4808]: I1002 17:09:58.879809 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb" Oct 02 17:09:59 crc kubenswrapper[4808]: I1002 17:09:59.437810 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb"] Oct 02 17:09:59 crc kubenswrapper[4808]: I1002 17:09:59.476707 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb" event={"ID":"7d9cd742-d1a8-48ae-bd84-3660eb151623","Type":"ContainerStarted","Data":"48c1e0e6097fa8786ad114dbf98cb0b0ba2dfba180410ee6b69ea156f65dffd4"} Oct 02 17:10:00 crc kubenswrapper[4808]: I1002 17:10:00.485111 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb" event={"ID":"7d9cd742-d1a8-48ae-bd84-3660eb151623","Type":"ContainerStarted","Data":"33793cb3670244a9fde4ac84af044a586bf2f660dca4cb721dedf604dec05395"} Oct 02 17:10:00 crc kubenswrapper[4808]: I1002 17:10:00.506644 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mxrqb" podStartSLOduration=2.051089907 podStartE2EDuration="2.506622648s" podCreationTimestamp="2025-10-02 17:09:58 +0000 UTC" firstStartedPulling="2025-10-02 17:09:59.44785619 +0000 UTC m=+1766.773385240" lastFinishedPulling="2025-10-02 17:09:59.903388981 +0000 UTC m=+1767.228917981" observedRunningTime="2025-10-02 17:10:00.50182438 +0000 UTC m=+1767.827353390" watchObservedRunningTime="2025-10-02 17:10:00.506622648 +0000 UTC m=+1767.832151658" Oct 02 17:10:08 crc kubenswrapper[4808]: I1002 17:10:08.396956 4808 scope.go:117] "RemoveContainer" containerID="a823e8cfe9d799300a4fab46e28dedc29e9fa6954d8fd2945bf7c266df1f72ba" Oct 02 17:10:08 crc kubenswrapper[4808]: E1002 17:10:08.397948 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7z66r_openshift-machine-config-operator(f83ce425-101d-4489-94a4-5c256eb29328)\"" pod="openshift-machine-config-operator/machine-config-daemon-7z66r" podUID="f83ce425-101d-4489-94a4-5c256eb29328" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515067530770024457 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015067530771017375 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015067525004016511 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015067525005015462 5ustar corecore